ESTeam

language technology software

  • Contact
  • Imprint
News
  • All
  • Meet us
  • Posts
Solutions
Activities
Company
28 Jan 2021
by Michael Wetzel

Semantix & ESTeam Win Giant EU Commission Contract

Semantix and its long-term partner ESTeam have been awarded the EU Commission’s call for tender GROW/2020/OP/0002. The service contract represents a total value of 34 MEUR and is one of the largest in Europe. Semantix is the leading language company in the Nordics, currently ranking #13 worldwide. ESTeam is a tech company known for innovative NLP, knowledge management, and multilingual data solutions. Further a significant partner for the collaboration will be Summa Linguae.

The contract is for translation services associated with the technical regulations and rules on information society services laid down in Directive (EU) 2015/15351. Also called the "Single Market Transparency Directive, SMTD", this procedure is a preventive monitoring instrument managed by the Directorate-General for Internal Market, Industry, Entrepreneurship and SME's for the whole Commission to ensure the proper functioning of the internal market.

The contract was won in competition with Europe’s most experienced services providers. It marks a breakthrough in offering an AI-driven solution for translation services that adds unified human and linguistic asset classification, content recycling, Coreon's knowledge-based terminology recognition, and semantic quality assessment to the process.

Britta Aagaard, Chief Business Officer at Semantix states: ”The award of this major contract and the innovative solution proposed will further position Semantix as industry leading language technology company for multilingual services”, and she continues, “For many years, ESTeam and Semantix have successfully collaborated on highly complex large-scale projects combining multilingual processing with a semantic information infrastructure and a powerful services organisation.”

Jochen Hummel, CEO of ESTeam: “In spite of the breath-taking progress of AI, customers of LSPs haven’t enjoyed much innovation since the rollout of CAT in the 90s. We are thrilled that in partnership with proactive Semantix we are now delivering the game-changing benefits of state-of-the-art translation solutions to important customers.”

Coreon
Summa Linguae
Semantix

10 May 2019
by Jochen Hummel

Sunsetting CAT

For decades Computer Assisted Translation based on translation memories has been the standard tool for going global. Although CAT had been originally designed with a mid-90s PC in mind and there have been proposals for changing the underlying data model, the basic architecture of CAT has been left unchanged. The dramatic advances in Neural Machine Translation (NMT) have now made the whole product category obsolete.

NMT Crossing the Rubicon

While selling translation memory I always said, machines will only be able to translate once they understand text. And if one day machines would, MT will be a mere footnote of a totally different revolution. Turns out that neural networks stacked deeply enough do understand us sufficiently to create a well formed translation. Over the last two years NMT has progressed dramatically. It has now achieved “human parity” for important language pairs and domains. That changes everything.

Industry Getting it Wrong

Most players in the $50b translation industry, service providers but also customers, think that NMT is just another source for a translation proposal. In order to preserve their established way of delivery they pitch the concept of “augmented translation”. However, if the machine translation is as good (or bad) as human translation, who would you have revise it, another translator or a subject matter expert?
Yes, the expert who knows what the text is about. The workflow is thus changing to automatic translation and expert revision. Translation becomes faster, cheaper, and better!

Different Actors, Different Tools

A revision UI will have to look very different to a CAT tool. The most dramatic change is that revision UI has to be extremely simple. To support the current model of augmented translation, CAT tools have become very powerful. However, their complexity can only be handled by a highly demanded group of ten thousands of professional translators globally.
For the new workflow a product design is required that can support dozens of millions of, mostly occasional, revisers. Also, the revisers need to be pointed to the texts which need revision. This requires multilingual knowledge.

Disruption Powered by Coreon

Coreon can answer the two key questions for using NMT in a professional translation workflow: a) which parts of the translated text are not fit-for-purpose and b) why not? To do so the multilingual knowledge system classifies linguistic assets, human resources, QA, and projects in a unified system which is expandable, dynamic, and provides fallback paths. In the future linguists will engineer localization workflows and create multilingual knowledge in Coreon. "Doing words” is left to NMT.

Blog post that originally occurred in Dec 2018 on the Multilingual Knowledge Blog

06 Apr 2018
by Michael Wetzel

Concept Maps Everywhere DTT Symposion 2018

On March 22-24 the DTT Symposion (short DTT) took place again in Mannheim. It is the bi-annual meeting of the German Terminology Association (Deutscher Terminologietag). We were exhibiting and I enjoyed talking to many Coreon customers there. It was a truly exciting event this year and according to the organizers the most busy ever. 200+ participants meant house full!

After a half day of pre-event workshops, the event kicked off Friday morning with a presentation from Martin Volk (University Zürich) on parallel corpora, terminology extraction, and MT. Martin challenged the hype around Neural Machine Translation and pinpointed some weaknesses: “NMT operates with a fixed vocabulary. But real world translation has to deal with new words constantly … how can we ensure terminology-consistent translation?”. His research confirms what we've outlined in an blog post: Why Machine Learning still Needs Humans for Language.

“Concept Maps Everywhere”

Back to the event … as one participant tweeted, concept maps were the dominating topic throughout the days. First a workshop by Annette Weilandt (eccenca) on taxonomy, thesauri, and ontologies, followed by a presentation by Petra Drewer (University Karlsruhe). Petra unveiled a plethora of benefits:

  • insight into the domain
  • systematic presentation
  • clear distinction between concepts
  • identification of gaps
  • equivalence checks across languages
  • new opportunities in AI contexts

No surprise, my event highlight was the Coreon customer presentation from Liebherr on the benefits of multilingual knowledge systems. In this very entertaining presentation Lukas Auer (Liebherr MCCtec) and Johannes Widmann (Liebherr Holding) outlined how pragmatic and effective the work with concept systems turns out. They concluded: “If we all think in networks, why should our termbase then be designed as an alphabetic list of terms??” Instead, the concept system driven approach has many advantages such as training of new staff, context knowledge for technical authors and translators, terminological elaboration of specific domains, insight into the degree of how far a domain is already covered, avoiding doublettes etc. Download a case study from the Coreon web site.

DTT 2018 Award for a Master Thesis on Coreon

And then the “i-Tüpfelchen” (cherry on the cake) on Friday afternoon: David Reininghaus received this year’s DTT award on his master thesis: “Applying concept maps onto terminology collections: implementation of WIPO terminology with Coreon”. David analyzed in his work how a real graph driven technology outperforms simple hyperlink based approaches: no redundancies, more efficient, less error-prone. David further developed an XSL-based method how to transform the MultiTerm / TBX hyperlink based workarounds into a real graph, visualized in Coreon.

Deutsche Bahn: Terminology-Driven AI Applications

Tom Winter (Deutsche Bahn and President of the DTT) illustrated in his session how terminology boosts AI applications. Through already simple synonym expansion the intranet search engines are now more meaningful (a search for the unofficial "Schaffner" now finds even documents where only the approved "Zugbegleiter" was used). Other applications are automatic pre-processing of incoming requests in a customer query-answering system or even improving Alexa driven speech interaction at ticket vending machines … who says terminology is still a niche application?

From Language to Knowledge

I am excited about the evolution of the DTT in recent years. How many more participants will we see in spring 2020? I am convinced the more the DTT community continues to leave the pure documentation niche and the more the focus moves onto areas that our customer Liebherr or Tom Winter have illustrated, the relevance and awareness level of the community will continue to grow. So that the organisers can again proudly announce:
Ausgebucht – no more seats left!

Original Blog Post from the Multilingal Knowledge Blog

12 Feb 2018
by Michael Wetzel

IoT Banks on Semantic Interoperability

The biggest challenge for widespread adoption of the Internet of Things is interoperability. A much-noticed McKinsey report states that achieving interoperability in IoT would unlock an additional 40% of value. This is not surprising since the IoT is in essence about connecting machines, devices, and sensors – ideally cross organization, cross industries, and even cross borders. But while technical and syntactic interoperability are pretty much solved, little has been available so far to make sure devices actually understand each other.

Focus Semantic Interoperability
Embedded Computing Design superbly describes the situation in a recent series of articles. Technical interoperability, the fundamental ability to exchange raw data (bits, frames, packets, messages), is well understood and standardized. Syntactic interoperability, the ability to exchange structured data, is supported by standard data formats such as XML and JSON. Core connectivity standards such as DDS or OPC-UA provide syntactic interoperability cross-industries by communicating through a proposed set of standardized gateways.
Semantic interoperability, though, requires that the meaning (context) of exchanged data is automatically and accurately interpreted. Several industry bodies have tried to implement semantic data models. However, these semantic data schemes have either been way too narrow for cross-industry use cases or had to stay too high-level. Without schemes data from IoT devices lack information to describe their own meaning. Therefore, a laborious and, worse, inflexible normalization effort is required before that data can be really used.
Luckily there is a solution: abstract metadata from devices by creating an IoT knowledge system.

Controlled Vocabulary and Ontologies
A controlled vocabulary is a collection of identifiers which ensure consistency of metadata terminology. These terms are used to label concepts (nodes) in a graph which provides a standardized classification for a particular domain. Such ontology, incorporating characteristics of a taxonomy and thesaurus, links concepts with their terms and attributes in semantic relationships. This way it provides metadata abstraction. It represents knowledge in machine-readable form and thus functions as a knowledge system for specific domains and their IoT applications.

IoT Knowledge Systems made Easy
A domain ontology can be maintained in a repository completely abstracted from any programming environment. It needs to be created and maintained by domain experts. With the explosive growth of IoT constantly new devices, applications, organizations, industries, and even countries are added. Metadata abstraction parallels object-oriented programming and unfortunately so do the tools used so far to maintain and extend ontologies.
But now the SaaS solution Coreon makes sure that IoT devices understand each other. Not only does Coreon function with its API as a semantic gateway in the IoT connectivity architecture, it also provides a modern, very easy-to-use application to maintain ontologies; featuring a user interface domain experts can actually work with. With Coreon they can deliver the knowledge necessary for semantic interoperability so that IoT applications can unlock their full value.

Bosch ConnectedWorld 2018 Coreon Presentation slides

18 Mar 2016
by Michael Wetzel

A Concept never Lives alone! DTT Symposion 2016, Mannheim, Germany

On 3 – 5 March the 15th DTT Symposion took place in Mannheim. With approximately 160 participants this bi-annual event is probably the largest German speaking conference on terminology. In an exhibition hall several tools and services providers including Coreon presented their offerings. The focus theme of this year's event was "Terminology and Culture". I had slight worries before the conference whether it will be too theoretic, with many "dry" lectures and without practical impact. But no – I was positively surprised by the opposite: Many presentations, particularly during the two afternoon sessions, made the bridge from the humanities research to best practices in daily life.

Focus on Concept Maps

In fact I was very happy to see that the topic of Concept Maps, that systematic terminology work via concept systems is now an established focus of the DTT. Many presenters support our approach and confirm the value of Multilingual Knowledge Systems. Below a couple of my, very personal, highlights and take-aways from the sessions:

Begriffsimperialismus

The first conference day morning sessions settled the theoretical foundations. Peter A. Schmitt's (University Leipzig) as well as Wolfgang Sturz' (Transline) contributions highlighted that you can't separate culture from technology. Namely when comparing two cultures one will see that a given artifact has in each culture very specific characteristics. This cultural context influences terminologies, for instance in a temporal way (past versus today) or regional (American – British).

A highlight for me was F. Massion's (D.O.G.) presentation on "Begriffsimperialismus" (can I translate this with "imperialism imposed through concepts"?), i.e. the questionable practice to rely on one language's concept system as the baseline for developing terminologies – Massion's most important point was to develop methods and tools that respect the different concept systems across organizations and languages. He dived deep into concept system modeling. I fully support the key message in his talk: A concept never lives alone! We learn new concepts via existing knowledge structures. A concept is a member of several concepts systems – or, in other words: as many experts, as many definitions! Therefore, tools that model concept systems must support polyhierarchy and it would be even better if they support multidimensional relations, as well as – ideally – allow role based views on the concept systems.

SAP's Cloud: Simplicity, Good Search, Little Training

Mark Childress (SAP, and also president of the DTT e.V.) illustrated SAP's vision of modernizing and moving the SAP terminology database into the cloud: users expect simplicity, simple search means, little training. The data must be accessible on every device, via keyboard or voice. And also support collaboration, i.e. involvement of external experts. For me Mark's presentation confirmed the product design strategy of Coreon – namely that not only consumer apps but also enterprise, B2B software solutions require appealing user interfaces together with intuitive interaction paths.

Pivoting at the EU

On the 2nd day of the conference, Rodolfo Maslias (European Parliament) broadened the view and the larger political context: the EU institutions' terminology work should support multicultural legislative procedures. How to tackle cultural differences (North-South, East-West)? Today, the EU has 24 official languages, but the "Europe of Nations" more and more evolves into a "Europe of Regions". Irish is currently being added to the list of languages, what would happen to languages such as Catalan? In daily life, the need to support 24*23 languages, i.e. 552 language combinations, is only fulfilled through the pivot languages English, German, French. Rodolfo also made a clear statement that a lingua franca such as English works only for basic needs. Whereas collaboration and expert exchange in each and every domain must allow the participants to use their mother tongue.

Merging Legal System in South Tyrol

Elena Chiocchetti (EURAC) then illustrated in the context of the South Tyrolian legal system the challenges to bring two concept systems (German, Italian) together. While in theory the goal is to identify equivalent and appropriate terms (and also a comparable government body) in real life it will be often rather an approximation driven by functional equivalency.

Multingual Knowledge with Coreon

I then had the pleasure to present just before lunch … I could nicely resume F. Massion's thoughts from the previous day: How ESTeam spin-off Coreon models different perspectives onto one and the same concept, what it means for its terms as well as its relations and how well the standards TBX and SKOS tackle these challenges.

The Must of Concept Maps

In a joint presentation Klaus Fleischmann (Kaleidoscope) and Alexandra Hanischläger (KTM) then launched a fireworks of arguments for working with concept maps: without a map it is difficult to identify terminological gaps – maps help to detect synonyms – concept relations express at least as much information as the written descriptions – union of textual plus semantical information helps to disambiguate. Last but not least, concept maps are very useful in training new employees. Anyone still questioning the benefit of multilingual knowledge systems?

Milos Jakubicek then illustrated "Sketch Engine", an online service for terminology extraction based on large corpora. While I can't yet judge about the linguistic quality of this service, I very appreciate the trend to move also terminology extraction into the cloud and to offer it as a hosted service.

I am looking forward to the 16th DTT Symposion in 2018 – by the way, Mannheim and the Dorint Hotel are a pretty good choice!

Conference Slides

15 Mar 2016
by Michael Wetzel

At NATO Headquarters: Terminology Experts Addressing Multilingual Knowledge NATO Terminology Conference, 19-20 November 2015

Last November 150 experts and I from all across the world attended the NATO Terminology Conference. The event took place at the campus of the NATO headquarters in Brussels, hosted by the NATO Standardization Office (NSO).

These two days were probably one of the most inspiring and valuable terminology conference days I've attended so far. I've never attended a terminology conference before where throughout the sessions the benefits of concept structures and multilingual concept maps were constantly emphasized. As far as I remember, it was Cristina Valentini from WIPO that cited one of her translators: "… only through the concept map did I understand what terminology is about".

I thank Folkert Zijlstra and his team from the NSO for a great conference, and also giving us the opportunity to present about the benefits of Multilingual Knowledge Systems. Below I want to give you a brief overview of what for me were the highlights at the conference.

Major General Edvardas Mazeikis from the Lithuanian Air Force and Director of NSO gave an excellent opening and setting the context for the conference with the theme "There is no interoperability without standardization".
Ian Jones, former head of the Linguistic Service at NATO SHAPE in Mons, continued to provide a highly interesting chronological overview and historical perspective on terminology work – from the beginning of human language during Stone Age to Johnson's dictionary from 1755 to Eugen Wüster's pioneer work and then to the first written terminology "cards" at NATO in 1956. Ian is also co-author of the recent book Meeting the Language Challenge of NATO Operations.
Folkert from the NSO outlined the internal NATO Terminology Programme. One of the most interesting aspects for me was to understand how an organization like NATO deploys a strict, multi-layered approval process: from Subject Matter Experts to Delegate Committees and Senior Committees.
In the following session Chantal Reid, Translation Bureau Government Canada, and Ian McIlroy, National Defence Canada, provided guidelines how to assign ownership: Standardize at the lowest level. As an example, terminology that is unique to one organization, may be approved by that organization, whereas terminology that spans more than one organization must be standardized at department level.

The afternoon sessions confirmed that conceptual structures are fundamental to properly manage terminology and that the linking of concepts takes terminology to a higher level, namely knowledge and when multilingual to a Multilingual Knowledge System.
Context, context, context! – Pamela Faber and Pilar León-Araúz from the University of Granada illustrated how principles of Frame Semantics are applied to terminology records, how knowledge structures, for instance levels of super-ordinate categories organize records: 'Patient transfer' -> 'evacuation' -> 'non-medical evacuation' -> 'casualty evacuation'. Pamela and Pilar highlighted how transparent this approach is and how it helps to fill gaps in the resource.
Cristina Valentini from the WIPO in Geneva, presented WIPO Pearl. I found it interesting to learn about the "paradox" when it comes to patent terminology: On the one hand the need for very precise terminology so that experts understand and assess a patent, but on the other hand the patent filer's interest who wants to achieve a very broad protection; therefore he chooses a less precise and rather vague terminology: compare 'glasses' – 'eyewear' versus the fuzzy construct 'eyesight correcting apparatus' … Noteworthy about Cristina's presentation for me was certainly the demo of WIPO Pearls concept cloud, WIPO's custom development of a concept map on top of existing terminology records. Through the concept map users trust the database more, since they become "confident in the unfamiliar".

Frieda Steurs, KU Leuven, started the 2nd day. She assessed whether ISO Terminology Standards are a fit for NATO. In this context – terminology standardization aims at avoiding synonymy and misunderstandings – I really liked one of Frieda's slides where she used the phrase "competing terms" talking about synonyms; I will use that from now on, too! Frieda also highlighted ISO 24156-1 and how UML diagrams are used to create concept maps, a good reference standard for tool developers.
Danielle Henripin from the United Nations shared best practices when it comes to unifying data from many different sources: flexibility of the future system, migration milestones that favor gradual progress, automate tedious tasks, provide guidelines and training to overcome reluctance in the user base, and always Listen, Learn, Adapt.

Conference Official Summary

12 Jun 2015
by Michael Wetzel

How can Language Technology Leverage Big Data and Transform the Localisation Industry? A Contribution to the GALAxy Newsletter

Jochen Hummel, ESTeam CEO and Chairman of LT Innovate, foresees a major disruption in the Localisation Industry in the near future — one that will elevate the question of multilingual data from translation processes to a C-level concern. How can Language Technology leverage Big Data and transform the industry? Within this framework, Hummel introduces this quarter's issue of the GALAxy newsletter and invites readers to engage in this brave new world.

GALAxy newsletter

05 Jun 2013
by Michael Wetzel

EDF 2013: Unlocking the Data Treasure Chest Video recording from the Language Technology Panel

The European Data Forum (EDF) 2013 took place from April 9-10, 2013 in Dublin, Ireland. EDF is the annual meeting-point for data practitioners from industry, research, the public-sector and the community, to discuss the opportunities and challenges of the emerging Big Data Economy in Europe. Gudrun Magnusdottir, CSO of ESTeam, had the pleasure to moderate the Language Technology Panel.

Language technologies are inevitable for adding value to Big Data. They are needed for extracting information and knowledge from textual documents and social media, and for making them accessible across language borders. But they are also required for searching, linking and documenting non linguistic, e.g. numerical or multi media data, through effective metadata. And they are the core of intuitive semantic interfaces to the human consumer of the end products, e.g., through question answering, report generation and summarization. The proposed joint panel by META-NET, LT Innovate and Multilingual Web is a call for action: to further explore the impact of language technologies on big data and to develop a plan – for leveraging big data to big knowledge.

View the complete recording including slides

03 Apr 2013
by Michael Wetzel

Mobility, Growth and Jobs - The Multilingual Challenge of the Single Market Event Report from Policy Dialogue, 5th Dec 2012 - Poliglotti4.eu

The EU’s multilingual nature can be an asset for European companies, giving them the tools they need to harness multilingualism as a means of accessing multicultural global markets and gaining a competitive  advantage over their US counterparts, heard participants in this Policy Dialogue, jointly organised by the EPC and the European Academy of Yuste Foundation, and with the kind support of Poliglotti4.eu project.

Jochen Hummel, Chairman of LT Innovate and CEO of ESTeam is quoted in this report, saying that while "… capital could flow freely between EU countries, the same could not be said of information, citing as an example the difficulties of cross-border e-government." And " …it’s not necessarily a translation problem. It’s a problem of processing multilingual information. The control of information and content has changed, shifting from companies to users and from governments to citizens"

European Academy of Yuste Foundation homepage
European Policy Center homepage
Poliglotti4.eu homepage
Full Event Report

15 Feb 2013
by Michael Wetzel

Eventful LISE Weeks - Valuing the Quality of Resources Workshop with IATE, Management and Advisory Meetings

Last Monday, 4th February, all LISE partners were on their way to Benelux.

The first event took place in Luxembourg: a joint workshop with IATE users from several European institutions (CdT, Court of Auditors, Parliament). Representatives from all LISE work packages teamed up with members from the IATE group to play with, evaluate and discuss LISE technology. ESTeam Tools – the technology in LISE – are now customized and enabled to process IATE data – with more than 11 million terms probably the largest European terminology collection. It turned out that particularly ESTeam Cleanup plays an indispensable role when assuring high quality terminology resources. High quality language resources – crucial in legal practices (LISE’s focus) as well as in the finance, insurance or health care business(es). One participant commented: “automatically spotting errors, misspellings, doublettes, wrong classification attributes … – this is a terminologist’s dream”.

We then all continued to Ghent to meet in the premises of LISE partner CrossLang. The official LISE EMB / TMB meetings carried on throughout Wednesday afternoon. Joeri van de Walle from CrossLang presented us with preliminary results from the IATE workshop, confirming the encouraging feedback from the day before. The final evaluation reports will contain qualitative (via a survey) as well as quantitative figures (via statistical capabilities in the ESTeam Tools). An additional workshop is already planned to take place in Brussels.

On Thursday we had the LISE Advisory Group meeting. Advisors from EMF, Swiss Federal Chancellery, and Infoterm assessed LISE, its technological and economic potential. Since the workshops have already confirmed the users’ needs and the value of the technology, we are now ready to focus on how to exploit and market the LISE value proposition. And, indeed, we are embracing several brilliant suggestions from the advisors on how to proceed.

Fascinating weeks!

LISE homepage

26 Jun 2012
by Michael Wetzel

Language Technologies are the Missing Piece in the Puzzle in Europe's Digital Agenda First LT-Innovate Summit 2012 in Brussels

On 19 June 2012, over 160 key players of the Language Technology (LT) industry met in Brussels for the first LT-Innovate Summit. The event showcased new technologies in LT, fostered investment opportunities, but most importantly, it successfully conveyed the message that Language Technologies are the missing piece in the puzzle of the EU’s strategy for the Digital Single Market.
Philippe Wacker, Secretary General of LT-Innovate (http://www.lt-innovate.eu), stated: “A multilingual infrastructure is as important for Europe as a broadband infrastructure! Making content produced in any language available to 500 million Europeans is the real opportunity in the decade ahead.”
Jochen Hummel, LT-Innovate Chairman and CEO of ESTeam: “The language infrastructure is an important piece of enabling technology that the EU should procure as soon as possible. It would be a major booster for the European content economy with positive spill-over effects for almost every segment of the economy and society at large.”

Event homepage
Full PR
LT-Innovate homepage

10 May 2012
by Michael Wetzel

ESTeam Signs 2-Year Framework Contract with European Commission

Berlin, Töreboda, May 2012 – The European Commission, Employment, Social Affairs and Inclusion DG, has signed a 2-year framework contract with ESTeam AB. In order to develop, manage and disseminate the ESCO classification (European skills/competences, qualifications, and occupations), the European Commission requires external expertise to support the development, enhancement, assessment and management of the ESCO taxonomy and other taxonomies.
ESCO’s objective is to develop a multilingual, structured, easy-to-use terminology to describe the most relevant skills, competences and qualifications needed for several thousand occupations. This brings benefits to both jobseekers and employers. For example, it could be used to help jobseekers better describe their skill sets, or to develop new training initiatives adapted to the needs of the labour market and improved career guidance services.

ESTeam is providing expertise on taxonomies, semantic technologies as well as controlled vocabularies and terminological resources.

EURES Portal (basis of the ESCO classification)

04 May 2012
by Michael Wetzel

ESTeam AB Wins 4-Year Framework Contract from OHIM

Berlin, Töreboda May 2012 – OHIM, the Office for Harmonization in the Internal Market (Trade Marks and Designs), based in Alicante, Spain, has signed a 4-Year framework contract with ESTeam AB. In a continuous effort to improve business processes, OHIM requires and implements high-quality and value-adding services with regard to the usage of domain language in legal decision support scenarios.

With many years of experience in intellectual property matters, ESTeam is delivering linguistic consultancy and development services, such as linguistic text mining and database cleaning, textual information retrieval, multilingual terminology management, resource providing and decision support analysis, tools and systems to OHIM.

OHIM homepage

09 Feb 2012
by Michael Wetzel

Jochen Hummel elected Chairman of LT-Innovate

In a support action the EU is funding the creation of a forum for language technologies (LT) companies. The Language Technology Innovation Forum had its constituting meeting on Jan 27 in Brussels. The forum aims to have hundred members by June this year. On June 20 it will have, adjacent to the META meeting in Brussels, its first General Assembly featuring the award session for the LT biz contest.

The forum is supposed to be driven by industry focusing on SMEs and their needs. Therefore the forum has created an Industrial Advisory Board and a Steering Committee. Initiatives shall be driven through Special Interest Groups. Jochen Hummel, CEO of ESTeam, has been elected chairman of the Steering Committee. He is heading the SIG for eGovernment aiming to position LT in the Semantic Layer of the EU’s Interoperability Model. He will also work closely with the forum on developing an Innovation Agenda for the LT industry.

Read more on lt-innovate.eu and follow @LTinnovate, @LangTechNews, and @JochenHummel on Twitter

LT-Innovate homepage

07 Feb 2012
by Michael Wetzel

LISE Meeting The first year

ESTeam was pleased to host the LISE EMB/TMB meetings in its offices in Berlin. On 2nd and 3rd February LISE consortium members from University Vienna, EURAC, the European Union, crosslang, and ESTeam met to review the first 12 months of the project and to plan the next steps and deliverables.

Project coordinator Tanja Wissik started with a status assessment. We particularly tried to identify items that require clarification before the first official project review meeting end of March with the EU project officer. We then listed and discussed upcoming marketing and dissemination activities, concretely the attendance of LISE partners at the IRIS 2012 conference end of February in Salzburg, the LREC in Istanbul and the TKE in Madrid. The last session of the first day was dedicated to the LISE business plan, where I’ve reported on ESTeam’s in-depth research on the terminology market landscape and how LISE could see commercial success.
On the second day Elena Chiocchetti from EURAC reported on their user research progress and about the outcome of interviews with terminology managers in administrations – the survey is still running and also available online. Then we’ve dedicated a lot of time to demonstrate the already fully functional LISE collaboration portal, developed by ESTeam. It was very encouraging to see that the user needs as identified by EURAC matches very well with what has been implemented in the portal. A fascinating project!

LISE homepage

25 Nov 2011
by Michael Wetzel

6th Ministerial eGovernment Conference A Review from Gudrun Magnusdottir

ESTeam attended the 6th European Ministerial eGovernment Conference, entitled "Borderless eGovernment Services for Europeans" on 17 – 18 November 2011 in the City of Poznań organized by the Polish Presidency of the European Union and the European Commission. The conference, a major event during the Polish Presidency, attracted approximately 900 stakeholders from around the world: ministers from EU member states, officials from national, regional and local administrations, industry representatives, civil society and experts.

The conference focused on three main aspects of eGovernance:

  • Policy issues
  • Interoperability strategies
  • People’s concerns for issues like the impact of eGovernment services, inclusiveness, public value, user empowerment and quality assurance.

Large scale projects such as Peppol were presented and they are due to end in less than one year from now. Regrettably only very few governments are implementing these projects into their infrastructure. The European Commission needs to take a greater responsibility for eGovernment projects that they fund and be the first to put them to use. The needs of the Commission must be considered in these projects since this is the only way that cross border operability will be achieved. The need for cross border services should be implemented by the owner of the problem, namely the organization that crosses EU borders.

Not a lot has happened since the Malmö conference during the last two years. We believe this is due to governments being reluctant to change but rather keep their own operational structure and cost. Furthermore it is apparent that implemented local solutions remain a functional island, missing the potential to be exploited further in other regions – not to mention even other countries. Interoperability and cross border services are fashionable terms in eGovernment but without more actual substance than the plain understanding that the problem is actually there. We are still far away from what to do about the problem.

It is noteworthy that no presented project was close to the scale of Euroclass. Euroclass supports SMEs across Europe to register a trademark in one single step in all EU languages and in any country in a pan European collaborative effort. In such areas progress rather looks slow, while every single step forward is extremely important, since it facilitates the communication between citizens and government as well as having the task to effectively cut the cost and improve the structure of government operations.

Words such as standardization and interoperability are frequently used. Question is how do you standardize in a world with such varied legal and parliamentary systems? Here you must decide to disagree and remember that the insight and knowledge about of where you disagree is also a result of harmonization and interoperability.

ESTeam met many fantastic people from all over Europe with great ideas. The human potential is clearly there for realizing solutions to change old fashioned management and communication methods in European governments.

Conference home page
Peppol home page
Euroclass

04 Nov 2011
by Michael Wetzel

CIKM Conference A Summary from Mihai Lupu, ESTeam

The Conference on Information and Knowledge Management (CIKM) has had its 20th anniversary this year in Glasgow. Over the years, CIKM developed into a prime meeting venue for all scientists interested in information technologies, from both academia and industry. This year, the conference accepted 137 of the 917 paper submissions it received (a 15% acceptance rate), and established its position as one of the largest and most selective conferences in the field.

In addition to all the academic papers, the Thursday industry session was particularly interesting. Stephen Robertson, one of the key personalities of information retrieval, started the session with an acknowledgement of the needs of professional search, in contrast to the vast majority of research done on web search. He argued that instead of viewing professional search as a specific case, it is the web search that is in fact the exception to the rule. Only in web search do we have a practically unlimited data source, where recall is rendered useless and precision rules. He mentioned legal, prior-art and healthcare search as the prototypical professional search scenarios, which require further attention from the information retrieval community. There were also other interesting talks, in particular that of Khalid Al-Kofahi (Thomson Reuters) describing a new search system for legal search applied to civil law jurisdictions.

Together with the main conference, a total of 15 workshops were organized, to cater to all the different aspects of information sources and uses. Among them, the Patent Information Retrieval (PaIR) workshop (4th edition) was organized by Mihai Lupu, member of ESTeam. PaIR 2011 was equally attended by both industry and academic participants. Among them, we counted Google, Springer and CambridgeIP.

The conference was not all about work and also invited the attendees to a series of social events, starting the night before the conference, with a whiskey distillery tour. It continued on Monday with a welcome reception, on Tuesday with a Civic Reception hosted by the Mayor of the city, and finally a conference banquet on Wednesday, in the stunningly-restored Kelvingrove Art Gallery and Museum.

CIKM home page
Patent Information Retrieval Workshop

30 Sep 2011
by Michael Wetzel

PLuTO Meeting Productive Sessions in Sunny Crete

ESTeam was pleased to host the EU PLuTO meeting in the very picturesque town of Chania in Crete. On 29th and 30th September PLuTO consortium members from Dublin City University, CrossLang, and ESTeam met to review the first 18 months of the project and to plan the next steps and deliverables.

We kicked off with a review from the PLuTO project coordinator John Tinsley. He particularly demoed a plug-in that makes patent translations available in web browsers upon a simple mouse-click. Joeri van de Walle from CrossLang then illustrated interesting facts about the quality of the machine translation results: the focus on the patent domain guarantees that PLuTO creates better results than other translation systems. Lambros Kranias from ESTeam proudly presented his research on the encouragingly high number of TM subsegment recycling rate. My presentation on potential strategies on commercialization and productization concluded the first day.

On the second day we’ve split into smaller groups and had several productive workshops spread all across the sunny conference site.

PLuTO homepage