Investors received multiple alerts about a security breach at Appen.
Appen Limited (ASX:APX), a multilingual AI enablement and machine learning company headquartered in Sydney, issued a report on July 30th (Australian time) that malicious actors had hacked a third-party provider and stolen access to its systems. Appen is listed on the Australian stock market, ASX.
The third-party system was being used on a trial basis and Appen ceased using it immediately. The company believes that the attack was random and not targeted specifically to its repositories, since other companies were also victims of similar incidents. Even though it was determined that the hackers gained access to Appen’s Annotation Platform, which contained customer and crowd names, company names, email addresses, encrypted passwords, IP addresses, and historical login and log off times, and some phone numbers, the IT security team asserts that the incident is limited in nature and not material.
According to the release issued to the Australian Stock Exchange, Appen — which analyzes data for eight of the largest ten technology companies in the world — has not suffered any interruption to its operations and has reported the incident to legal authorities.
The statement affirms that customer AI training data, the core business of the company, is stored separately and there is no evidence that it was stolen. The unauthorized access was detected as soon as it occurred, and Appen took the necessary steps to secure its systems. Relevant clients were contacted and had their passwords and security tokens reset. A cyber forensics firm was hired to assist in the investigation.
According to Dow Jones, Appen shares were up 1.2% at AUD 36.11. After a previous filing announcement before the breach, shares had been up as much as 3%.
Marketing games for international venues, celebrating the Lunar New Year with Chinese characters on jerseys, or recruiting a global class of players, the National Basketball Association is no stranger to localizing its content. But deepening AI and machine learning technologies promise an approach to its global fan base like never before.
The NBA announced in April that it has made plans to enter a multi-year partnership with Microsoft to create a more personalized, localized experience for its international fan base. But while fans around the world already enjoy watching NBA broadcasts in 47 languages broadcast in 215 countries, the partnership promises fundamental innovations to modernize fan interaction using artificial intelligence and machine learning technology.
One of the ways the alliance will change NBA content is by creating a direct-to-consumer platform on Microsoft Azure, Microsoft’s cloud computing service, which will provide data analytics, computing, storage, and networking that they anticipate will allow them to personalize fan experience “through state-of-the-art machine learning, cognitive search and advanced data analytics solutions,” according to the report.
Analyzing metrics around fan behavior is nothing new for the NBA, but with such data at its fingertips, the NBA-fan relationship may flourish like never before. NBA senior vice-president of direct to consumer Chris Benyarko said about the benefits of machine learning AI curating fan experience, “Instead of the fan having to pick and choose and turn them on or off one by one, the platform is now starting to behave like a game producer, automatically selecting and presenting the game in a different way.”
Deb Cupp, Corporate Vice President of Enterprise and Commercial Industries at Microsoft, said, “The AI eventually learns that I like to learn about stats, so it’s going to start presenting me more information about stats as I go into the game… It’s this experience where instead of just watching a game, it actually has the opportunity to interact in a way that matters to me as that fan.”
With fans in all different time zones who experience the game and consume content in a variety of ways, the NBA will be able to gather a global array of data to bring in new fans and retain old ones. What that will mean for sports remains to be seen, but the machine learning will likely provide invaluable information on fan behavior worldwide and grant the NBA a chance to solidify its place as a premier global sport.
Jonathan Pyner is a poet, freelance writer, and translator. He has worked as an educator for nearly a decade in the US and Taiwan, and he recently completed a master’s of fine arts in creative writing.
Let’s look over the past year to get a sense of how much progress we have made, and what progress may lie before us. AI has become a norm in our lives and is now part of the vernacular more than ever. People have accepted that their lives interact multiple times a day with AI, and that such technology is becoming ubiquitous.
What does 2019 hold for us? Well, if we pay attention to the predictions of 16th-century seer Nostradamus, we should brace ourselves for flooding, wars and a strike by a meteor. Nothing all that cheerful but, assuming we survive all of that, what does 2019 hold for technology trends?
We’ve seen an explosion in the use of AI in the delivery of neural machine translation during 2018; expect this to continue in 2019 and beyond. AI is the catch-all term to cover machine learning (ML) and deep learning. Machine learning is an over-arching term for the training of computers, using algorithms, to parse data, learn from it and make informed decisions based on accrued learning. Examples of ML in action is Netflix showing you what you might want to watch next. Or Amazon suggesting books you might want to buy.
Within the localization industry, the use of AI in the form of machine translation (MT) in several forms has significantly improved translation quality outputs, sped up translation of huge quantities of data and reduced the price of translation to make it economically viable.
AI refers to computer systems built to mimic human neural abilities and to perform tasks such as image recognition, parsing speech forms, discerning patterns from complex data sets and informing accurate decision-making. What’s more, AI can do these tasks faster, cheaper and more accurately than humans. Although AI has been around since the 1950s, it can be truly said that it has now come of age. This maturity has been propelled by the ever-increasing computational power now available in the cloud.
According to Forbes, five out of six people use AI technology each day. These services include such things as navigation apps, streaming services (such as Netflix), smartphone personal assistants, dating apps and even smart home devices (such as remote-activated home security systems). Additionally, AI is used in recommendation engines used by eCommerce sites to schedule trains, to predict maintenance cycles and for other mission-critical business tasks.
For the localization industry, AI will become a highly-integrated component of MT systems. The role of the human translator will continue evolving to that of an editor of MT texts, rather than translator of raw texts. In addition, pricing models will continue to move from the traditional price per word based on word volumes to pricing on a time-measured rate. MT will become an integral part of the standard workflow. The reality of real-time translation — driven by such technology as the internet of things (IOT) — will see project managers and editors managing workflows of projects required by customers who need a constant flow of updated information. MT will become part of the translation process just as much as other computer-aided translation tools did in the past. And, as ever, evolving technology will bring with it a desire for speedier and cost-effective solutions.
ML will continue to grow as a tool used by most localization departments as the requirement for the speedy translations of large datasets continues to be a driver in the industry.
ML is a subset of AI: with ML, computers are automated to learn to do something that they are not initially programmed to do. So, ML is an over-arching term for the training of computers to use smart algorithms to automate actions, to parse complex data and to learn patterns from this learning thus enabling the computer to make informed decisions based on this accrued knowledge. ML can be broadly broken down into two types of learning: supervised and non-supervised learning.
For supervised machine learning, the training data is pre-labelled and consists of an aligned input data set and desired output data set. For example, an input data set could be a translation memory. An ML algorithm analyses the training data and maps how to convert future inputs to match the learned, desired output data sets.
Unsupervised ML is like supervised ML; however, the input data sets are not pre-classified or labelled. The goal of unsupervised machine learning is to find hidden structures in the unlabelled data.
So how does this impact the localization industry? Well, suppose you want to build a translation system to translate from Zulu to French, without any Zulu-French training data. The solution is, you can combine both supervised and unsupervised approaches to achieve this. You can use an English-Zulu data set in combination with an English-French data set, and using unsupervised machine learning, the system can learn how to translate from Zulu into French.
This approach is commonly referred to as “zero-shot” machine learning — expect to hear more about this in 2019 for MT systems for long-tail languages.
While blockchain is most widely known as the technology behind cryptocurrencies, it offers security that is useful in many other ways.
In simple terms, blockchain can be described as data you can add to, but not take away from or change. These blocks of data can be “chained” together to create incredible secure data repositories. Not being able to change any previous blocks is what makes it so secure.
This enhanced security is why blockchain is used for cryptocurrencies. It is also why it will play a significant role in localization where it will used to protect information such as a client’s financial details, and to protect and preserve translation memories; especially in translation memories used in distributed translation workflow scenarios.
Cloud computing has now become mainstream: most of all global companies now rely on this centralized hosting structure for machine learning and powerful computational power. This cloud market is dominated by just a few gigantic companies such as Amazon, Microsoft, Google and IBM. However, now that we’ve been using cloud computing for some time, companies have realized that accessing all data from a central repository introduces a time-delay latency, which in turn slows down the delivery of services which can, in turn, increase costs. The “round trip” made by cloud-based data is seen by many of today’s companies as a hindrance to their business growth.
Technology stands still for no man, and so, for many, the cloud has reached its peak as a service for some technologies. The cloud will continue to be used to analyze and process huge swathes of data, but the advent of the IoT (connected security systems, electronic appliances, vending machines, automated lighting), where data processing needs to be high speed, if not real time, demands a different model. So the logical and necessary next move is to move this data processing to the edge. The edge simply means that data processing is moving from a far-away storage location to a geographical site closer to the data source. The advent of powerful computer chips that allows such processing to be done locally has expedited this move to the edge. Indeed, many of today’s cloud setups automatically look to place the processing of data at the optimum edge site for that data’s requirements.
So, edge computing solves the latency problem by simply moving the data processing closer to home. Closer to home means less time spent uploading and downloading data. Instead of the centralized storage model, which has hitherto driven AI, companies are moving their data into the “local community” to be processed. This move will undoubtedly make data access much faster and facilitate the growing demand for real-time computing.
How will this impact localization? Well, in 2019 we can expect to see the edge model used in domain-adapted MT systems, and distributed translation workflows that are designed to meet the increasing demand for data distribution in real-time.
We are on the verge of an explosion in the use of AI. The inevitable growth of AI will fundamentally re-shape how companies manage translation workflows; the very engine of their work process. Real-time translations will often become the norm.
I also predict that changes will happen at a human level; for example, the role of the translator will change from that of translator of raw text to that of editor of huge volumes of high-quality MT-produced text. This will be a beneficial change, allowing translators to increase their capacity and so increase their income. In 2019, we predict that the overall transformation effected by the advent of AI at all levels of the industry will bring with it an increased velocity of production, an improved efficiency in the delivery of translations, and a reduction in the cost of translating huge volumes of data.
Tony O’Dowd is the founder and chief architect of KantanMT.com, a cloud-based custom machine translation solutions provider. He is a serial entrepreneur and localization veteran, with almost 30 years’ experience working in the localization industry.
Privacy & Cookies Policy
Necessary cookies are absolutely essential for the website to function properly. This category only includes cookies that ensures basic functionalities and security features of the website. These cookies do not store any personal information.
Any cookies that may not be particularly necessary for the website to function and is used specifically to collect user personal data via analytics, ads, other embedded contents are termed as non-necessary cookies. It is mandatory to procure user consent prior to running these cookies on your website.