Click here to check if anything new just came in.
September 26 2014
September 25 2014
The next CKAN Association Community meeting is planned for this Friday 26th September at 4pm London / 11am EDT.
One major proposed topic is the Roadmap but please suggest additional topics in the meeting doc or here on list.
To help us manage the meeting and get an idea of numbers if you intend to come please add yourself to the participants list in the meeting doc.
- When: Friday 26 Sepember 2014 – 4pm London (BST) / 11am US East Coast (EDT) / 5pm European (CET)
- Where: online (likely via WebEx or similar – link to come)
- Text chat: http://webchat.freenode.net/?channels=ckan
- Chair: Jeanne Holm (GSA / Data.Gov)
- Meeting Doc: Google Doc here – please add your name if you plan to come!
- CKAN Roadmap
- Objectives and Priorities
- Communications and coordination
- Suggest topics here …
September 24 2014
September 23 2014
Today, we are thrilled to announce a partnership with Getty Images, the global leader in digital media, to offer Zemanta users the ability to embed high quality images directly from our Editorial Assistant. Whether it’s improving SEO or helping authors tell a better story, images play a critical role in the content creation process. The use of high quality images can help your blog stand above the rest. Now with Zemanta Editorial Assistant, bloggers can draw from a library of more than 50 million free images to compliment their content. Let’s take a look at how to embed Getty Images using Editorial Assistant. Just like our existing image recommendations, images from the Getty Images library are available via the Media Gallery.
- As you begin writing, Editorial Assistant will make image recommendations based on the topic of your post. Draw on Getty Images’ latest news, sports, celebrity, music and fashion coverage; immense digital photo archive; and rich conceptual images to illustrate your unique passions, ideas and interests. This partnership opens one of the largest, deepest and most comprehensive image collections in the world for easy sharing.
- Hover over the images for larger view, and click on an image once you are ready to embed it in an article. The image itself will be inserted to the position of your pointer in the text editor, but it’s also worth noting that the image is one-size-fits-all, which should fit all WordPress’s design layouts. It really is that easy.
- Also — all images already have their credits with the source URL hardcoded at the very bottom of the image, so that you won’t have to worry about adding it later.
- Once you’ve selected an image, make sure to include some in-text links and related articles recommended by Editorial Assistant to make your post even richer.
Getty Images embeds are available via both our Editorial Assistant WordPress plugin and Browser Extension. If you haven’t tried Editorial Assistant just yet, or are looking to start blogging again, you can download the plugin or browser extension here. If you have any question or concerns getting setup, we’re always happy to help. Send us an email at email@example.com, or find us on Twitter @Zemanta.
From 13th to 15th August 2014, the 15th IEEE International Conference on Information Reuse and Integration took place in San Francisco, USA. Sabrina Neururer had the chance to present the accepted paper entitled “Towards a Technology Roadmap for Big Data Applications in the Healthcare Domain” to a broad public on the first day of the conference. Very good feedback was received that confirmed the results of the presented study and highlighted the huge impact of Big Data applications on the healthcare domain. Technical challenges, such as semantic annotation, data sharing, data quality, privacy and security, and open R&D questions were discussed. Also non-technical challenges for Big Data Applications, such as user acceptance, were debated intensively.
September 19 2014
Big Data Project Final Event Workshop
In this workshop we will present the results of the BIG project including analysis of foundational Big Data research technologies, technology and strategy roadmaps to enable business to understand the potential of Big Data technologies across different sectors, and the necessary collaboration and dissemination infrastructure to link technology suppliers, integrators and leading user organizations.
PPP approved: the Big Project is pleased to have achieved its goal of creating a Big Data Public Private Partnership (PPP). The PPP will be signed by Neelie Kroes on October 13th and will allow industry to influence Europe's funding policy. Learn more about it in Session 2 of our workshop
The workshop is highly interactive, if you wish to give a position statement, please contact firstname.lastname@example.org (please include [finalevent] in the subject
Registration is still open (and free) at our workshop website
Date: September 30, 2014
Time: 9 a.m.–5:30 p.m.
Venue: Heidelberg Marriott Hotel, Heidelberg, Germany
Co-located with: ISC BigData 2014
The workshop is free to attend; lunch and refreshments will be provided by the BIG Project. In case of questions regarding the registration for the event email to email@example.com (please include [finalevent] in the subject of your email for faster response).
Four years ago we were standing in the offices of DG INSFO, the European Commission ministry of information and networks currently called DG Connect, for a first meeting on what is today the Digital Agenda Scoreboard.
DG INFSO invited us to discuss the future of the digital agenda scoreboard. Their objective was: create together with us a Linked Open Data platform for publishing statistics on how digital Europe is.
We did the classical data extraction towards Linked Data turning the CSV representation of the data into RDF according to the Data Cube Vocabulary. This vocabulary was then brand new and - as far as we know - we were one of the first to build an application with it. A lot of effort went into the creation of the use case scenario's, like:
- Visitors can select the data dimensions of their interest.
- The system creates dynamically graphs with all the background information attached.
Already at the first meeting we pointed out the new scoreboards potential for triggering vivid discussions on the data. This potential was shown right before the launch of the first version. During the preview at DG Connect, we showed a graph, and while explaining the selecting a vivid discussion on the correctness of the data started between the business owners. Mission accomplished, I thought.
Now 4 years later, my twitter feed shows on a regular basis tweets pointing to the scoreboard. Boring statistics in a book have been turned in a lively, public place where European citizens discuss on the how digital their country is.
Looking back, I am proud to be part of this story. It shows that breaking down the walls around data can kick start discussions in a community. Also the Digital Agenda Scoreboard was the start of the development of cool extensions to deal with publishing statistical data as Linked Open Data within the LOD2 project.
September 15 2014
At Zemanta, the tools and services we build are designed to help content creators create and then reach their total potential audience. Independent bloggers, publishers, and Fortune 500 brands work with Zemanta to create and amplify high quality content. To support our customers’ goals we rely on a diverse partner ecosystem — from Advertising Networks to WordPress Hosts.
If you run a network, platform, or software service that is seeking to further monetize or add a valuable service, let’s chat! Take a look at the opportunities we have available:
- Advertising Networks. Zemanta is the first dedicated Demand Side Platform (DSP) for Content Ads. Partner with Zemanta to monetize your network by accessing our inventory of high quality content ads programmatically. Our unique Promoted Content API enables Ad Networks full control over their content ads. Promoted Content from Zemanta can be found on networks such as Outbrain, Yahoo!, nRelate, and many many others.
- WordPress Hosts. Give your WordPress users the power of our Editorial Assistant WordPress plugin by default to create richer posts more efficiently and share in the revenues we create with our unique Editorial Network ad product.
- Content Creation, Curation, and Marketing Platforms. Partner with Zemanta to monetize your “authors” as an audience with Editorial Assistant or enable your customers to easily amplify their original content via our Content Ad DSP.
Do you run a platform or site that would benefit from a partnership with Zemanta? Visit our new partner page or email us at firstname.lastname@example.org to learn more. Even if one of the opportunities above doesn’t quite fit with your projects we’d still love to hear from you.
September 10 2014
September 09 2014
At September 1st at the 25th European Medical Informatics Conference (MIE 2014), Dr. Sonja Zillner presented "User Needs and Requirements Analysis for Big Data Healthcare Applications" which was received very well . As the realization of the promising opportunities of big data technolgies for healthcare relies on the integrated view on heterogeneoaus health data sources in high quality and the availablitlity of legal frameworks for secure data sharing, an intensive discussion of how to address the mentioned health data management challenges was triggered.
we are happy to announce the release of DBpedia 2014.
The most important improvements of the new release compared to DBpedia 3.9 are:
1. the new release is based on updated Wikipedia dumps dating from April / May 2014 (the 3.9 release was based on dumps from March / April 2013), leading to an overall increase of the number of things described in the English edition from 4.26 to 4.58 million things.
2. the DBpedia ontology is enlarged and the number of infobox to ontology mappings has risen, leading to richer and cleaner data.
The English version of the DBpedia knowledge base currently describes 4.58 million things, out of which 4.22 million are classified in a consistent ontology (http://wiki.dbpedia.org/Ontology2014), including 1,445,000 persons, 735,000 places (including 478,000 populated places), 411,000 creative works (including 123,000 music albums, 87,000 films and 19,000 video games), 241,000 organizations (including 58,000 companies and 49,000 educational institutions), 251,000 species and 6,000 diseases.
We provide localized versions of DBpedia in 125 languages. All these versions together describe 38.3 million things, out of which 23.8 million are localized descriptions of things that also exist in the English version of DBpedia. The full DBpedia data set features 38 million labels and abstracts in 125 different languages, 25.2 million links to images and 29.8 million links to external web pages; 80.9 million links to Wikipedia categories, and 41.2 million links to YAGO categories. DBpedia is connected with other Linked Datasets by around 50 million RDF links.
Altogether the DBpedia 2014 release consists of 3 billion pieces of information (RDF triples) out of which 580 million were extracted from the English edition of Wikipedia, 2.46 billion were extracted from other language editions.
Detailed statistics about the DBpedia data sets in 28 popular languages are provided at Dataset Statistics page ( http://wiki.dbpedia.org/Datasets2014/DatasetStatistics ).
The main changes between DBpedia 3.9 and 2014 are described below. For additional, more detailed information please refer to the DBpedia Change Log ( http://wiki.dbpedia.org/Changelog ).
1. Enlarged Ontology
The DBpedia community added new classes and properties to the DBpedia ontology via the mappings wiki. The DBpedia 2014 ontology encompasses
- 685 classes (DBpedia 3.9: 529)
- 1,079 object properties (DBpedia 3.9: 927)
- 1,600 datatype properties (DBpedia 3.9: 1,290)
- 116 specialized datatype properties (DBpedia 3.9: 116)
- 47 owl:equivalentClass and 35 owl:equivalentProperty mappings to http://schema.org
2. Additional Infobox to Ontology Mappings
The editors community of the mappings wiki also defined many new mappings from Wikipedia templates to DBpedia classes. For the DBpedia 2014 extraction, we used 4,339 mappings (DBpedia 3.9: 3,177 mappings), which are distributed as follows over the languages covered in the release.
- English: 586 mappings
- Dutch: 469 mappings
- Serbian: 450 mappings
- Polish: 383 mappings
- German: 295 mappings
- Greek: 281 mappings
- French: 221 mappings
- Portuguese: 211 mappings
- Slovenian: 170 mappings
- Korean: 148 mappings
- Spanish: 137 mappings
- Italian: 125 mappings
- Belarusian: 125 mappings
- Hungarian: 111 mappings
- Turkish: 91 mappings
- Japanese: 81 mappings
- Czech: 66 mappings
- Bulgarian: 61 mappings
- Indonesian: 59 mappings
- Catalan: 52 mappings
- Arabic: 52 mappings
- Russian: 48 mappings
- Basque: 37 mappings
- Croatian: 36 mappings
- Irish: 17 mappings
- Wiki-Commons: 12 mappings
- Welsh: 7 mappings
- Bengali: 6 mappings
- Slovak: 2 Mappings
3. Extended Type System to cover Articles without Infobox
Until the DBpedia 3.8 release, a concept was only assigned a type (like person or place) if the corresponding Wikipedia article contains an infobox indicating this type. Starting from the 3.9 release, we provide type statements for articles without infobox that are inferred based on the link structure within the DBpedia knowledge base using the algorithm described in Paulheim/Bizer 2014 ( http://www.heikopaulheim.com/documents/ijswis_2014.pdf ). For the new release, an improved version of the algorithm was run to produce type information for 400,000 things that were formerly not typed. A similar algorithm (presented in the same paper) was used to identify and remove potentially wrong statements from the knowledge base.
4. New and updated RDF Links into External Data Sources
We updated the following RDF link sets pointing at other Linked Data sources: Freebase, Wikidata, Geonames and GADM. For an overview about all data sets that are interlinked from DBpedia please refer to http://wiki.dbpedia.org/Interlinking .
Accessing the DBpedia 2014 Release
You can download the new DBpedia datasets from http://wiki.dbpedia.org/Downloads .
As usual, the new dataset is also available as Linked Data and via the DBpedia SPARQL endpoint at http://dbpedia.org/sparql .
Lots of thanks to
- Daniel Fleischhacker (University of Mannheim) and Volha Bryl (University of Mannheim) for improving the DBpedia extraction framework, for extracting the DBpedia 2014 data sets for all 125 languages, for generating the updated RDF links to external data sets, and for generating the statistics about the new release.
- All editors that contributed to the DBpedia ontology mappings via the Mappings Wiki.
- The whole DBpedia Internationalization Committee for pushing the DBpedia internationalization forward.
- Dimitris Kontokostas (University of Leipzig) for improving the DBpedia extraction framework and loading the new release onto the DBpedia download server in Leipzig.
- Heiko Paulheim (University of Mannheim) for re-running his algorithm to generate additional type statements for formerly untyped resources and identify and removed wrong statements.
- Petar Ristoski (University of Mannheim) for generating the updated links pointing at the GADM database of Global Administrative Areas. Petar will also generate an updated release of DBpedia as Tables soon.
- Aldo Gangemi (LIPN University, France & ISTC-CNR, Italy) for providing the links from DOLCE to DBpedia ontology.
- Kingsley Idehen, Patrick van Kleef, and Mitko Iliev (all OpenLink Software) for loading the new data set into the Virtuoso instance that serves the Linked Data view and SPARQL endpoint.
- OpenLink Software (http://www.openlinksw.com/) altogether for providing the server infrastructure for DBpedia.
- Michael Moore (University of Waterloo, as an intern at the University of Mannheim) for implementing the anchor text extractor and and contribution to the statistics scripts.
- Ali Ismayilov (University of Bonn) for implementing Wikidata extraction, on which the interlanguage link generation was based.
- Gaurav Vaidya (University of Colorado Boulder) for implementing and running Wikimedia Commons extraction.
- Andrea Di Menna, Jona Christopher Sahnwaldt, Julien Cojan, Julien Plu, Nilesh Chakraborty and others who contributed improvements to the DBpedia extraction framework via the source code repository on GitHub.
- All GSoC mentors and students for working directly or indirectly on this release: https://github.com/dbpedia/extraction-framework/graphs/contributors
The work on the DBpedia 2014 release was financially supported by the European Commission through the project LOD2 - Creating Knowledge out of Linked Data (http://lod2.eu/).
Have fun with the new DBpedia 2014 release!
Daniel Fleischhacker, Volha Bryl, and Christian Bizer
September 08 2014
SEMANTiCS conference celebrated its 10th anniversary this September in Leipzig. And this year’s venue has been capable of opening a new age for the Semantic Web in Europe - a marketplace for the next generation of semantic technologies was born.
As Phil Archer stated in his key note, the Semantic Web is now mature, and academia and industry can be proud of the achievements so far. And exactly that fact gave the thread for the conference: Real world use cases demonstrated by industry representatives, new and already running applied projects presented by the leading consortia in the field and a vivid academia showing the next ideas and developments in the field. So this years SEMANTiCS conference brought together the European Community in Semantic Web Technology – both from academia and industry.
- Papers and Presentations: 45 (50% of them industry talks)
- Posters: 10 (out of 22)
- A marketplace with 11 permanent booths
- Presented Vocabularies at the 1st Vocabulary Carnival: 24
- Attendance: 225
- Geographic Coverage: 21 countries
This year’s SEMANTiCS was co-located and connected with a couple of other related events, like the German ISKO, the Multilingual Linked Open Data for Enterprises (MLODE 2014) and the 2nd DBpedia Community Meeting 2014. This wisely connected gatherings brought people together and allowed transdisciplinary exchange.
Recapitulatory speaking: This SEMANTiCS has opened up new sights on Semantic Technologies, when it comes to
- industry use
- problem solving capacity
- next generation development
- knowledge about top companies, institutes and people in the sector
- Save the date for SEMANTiCS 2015: 15th – 17th of September 2015, Vienna
- SEMANTiCS 2014 – picture gallery: Flickr
We are delighted to announce that we have finalized the initial membership of the CKAN Association Steering Group. At present it consists of 4 organizations and their representatives as follows:
- Antonio Acuña (*), Head of Data.gov.uk, UK Cabinet Office
- Jeanne Holm, Evangelist, Data.gov, U.S. General Services Administration (Data.gov)
- Pat McDermott and Ashley Casovan, Open Government Secretariat, Treasury Board of Canada (data.gc.ca)
- Rufus Pollock (+), President (Open Knowledge)
(*) indicates Chair of the Steering Group
(+) indicates Secretary of the Steering Group
Steering Board Members will serve an initial term of 2 years and serve “ex officio” representing their organizations. We will continue to review Steering Group Membership and to consider potential new members as we go forward.
More Information About the Steering Group
The Steering Group is a key part of the CKAN Association. It is made up of key stakeholders who have committed to oversee and steer the CKAN Association going forward. The initial selection of the steering committee was coordinated by Open Knowledge.
Full details of the Steering Group can be found on the Steering Group page.
September 07 2014
PoolParty Server 4.5 as a major release of the PoolParty suite is available now. Learn more about the new features and improvements:
Automatic Batch Linking between Taxonomies
For linked projects a batch linking mechanism is available. The mechanism allows to calculate possible links between whole projects or subtrees of projects.
Create SKOS-XL labels with PoolParty
PoolParty has been extended to provide the ability to create SKOS-XL labels for concepts and add custom relations and attributes for SKOS-XL labels.
Improved Free Terms suggestion in Corpus management
A quality measure has been integrated into corpus management to show the quality of suggested terms.
Extended MS Excel import
Excel import now covers all SKOS properties and includes also all custom schema properties.
Find all new features, improvements and bug fixes in our release notes.
September 01 2014
With the arrival of semantic web standards and linked data technologies, new options for smarter content management and semantic search have become available. Taxonomies and metadata management shall play a central role in your content management system: By combining text mining algorithms with taxonomies and knowledge graphs from the web a more accurate annotation and categorization of documents and more complex queries over text-oriented repositories like SharePoint, Drupal, or Confluence are now possible.
Nevertheless, the predominant opinion that taxonomy management is a tedious process currently impedes a widespread implementation of professional metadata strategies.
In this webinar, key people from the Semantic Web Company will describe how content management and collaboration systems like SharePoint, Drupal or Confluence can benefit from professional taxonomy management. We will also discuss why taxonomy management is not necessarily a tedious process when well integrated into content management workflows.
In this webinar, you will learn about the following questions and topics:
… how standards like SKOS build the foundation of enterprise taxonomies to be linked with and enriched by additional vocabularies and taxonomies
… how knowledge graphs can be used to build the backbone of metadata services in organizations
… how automatic text mining can be used to create high-quality taxonomies and thesauri
… how automatic tagging can be integrated into existing content workflows of your SharePoint, Drupal or Confluence system
… how search-driven applications will become feasible when accurate metadata becomes available
Based on PoolParty Semantic Platform, you will see several live demos of end-user applications based on taxonomies and linked data.
We will showcase PoolParty’s latest release which provides outstanding facilities for professional linked data management, including entity extraction, PowerTagging, corpus analysis, taxonomy, thesaurus and ontology management.
August 29 2014
Maybe Soup is currently being updated? I'll try again automatically in a few seconds...