How Adobe, Dallas Zoo and Financial Fabric Are Using Microsoft Cognitive Services and the Azure Data Platform

A montage of recent customer use cases, demonstrating the range of capabilities of Microsoft’s AI, Big Data & Advanced Analytics offerings.

Adobe Helps Customers Break the Language Barrier

Adobe Experience Manager (AEM) is an enterprise content management platform that helps large and mid-size companies manage their web content, digital assets, online communities and more. Digital marketers use AEM to manage and personalize online content, including user-generated content on community sites. AEM helps companies build their brands online, driving demand and expanding their markets.

Companies using AEM often serve customers in many countries and languages, and customers were looking to Adobe for an easy way for their digital marketers to translate multinational web sites and content from international user communities into other languages. The sheer volume of information being posted online and prohibitive costs of human translation services meant that most companies could not afford the time and budget needed for manual translation efforts.

adobe-experience-manager

Adobe decided to address the problem by integrating AEM with Microsoft Translator. The Microsoft research team behind the Translator technology had spent over a decade to build a linguistically informed, statistical machine translation service that learns from prior translation efforts. What’s more, the service was flexible, reliable and performant at a massive scale, having performed billions of translations daily, and all while respecting customer data privacy.

To make the system work, all that Adobe’s customers had to do was supply the system with a set of parallel documents, containing the same information in two languages (i.e. the source and target languages). The system would then analyze this material and build a statistical model layered on top of Translator’s generic language knowledge. Translator’s statistical models and algorithms allow the system to automatically detect correlations between source and target language in the training data, helping it determine the best translation of a new input sentence.


AEM versions 6.0 and above now ship with a pre-configured Translator connector, and include a free trial license of 2 million characters per month, enabling users to start automatic translation with minimal effort. Adobe customers can now build, train, and deploy customized translation systems that understand a preferred terminology or style that is specific to their industry or domain. Some customers are also using this new capability to further expand their audience, providing content in previously underserved languages where there is a latent demand or audience. You can learn more about the Adobe solution here.

As people all over the world increasingly look to online communities for instant answers and information, they do not want to be restricted by their language skills. By partnering with Microsoft, Adobe is helping customers become productive by breaking through language barriers.

Dallas Zoo Tracks Elephant Behavior with Fitness Bands

Zoos worldwide have been working hard to provide larger, more natural and varied environments for their elephants. The Dallas Zoo, however, decided to go an extra step further: A growing number of elephants at the Dallas Zoo are now part of a pioneering application that uses RFID bracelets and Microsoft Azure to better understand elephant behavior and provide more customized care.

The earlier solution deployed by the Zoo relied on a combination of video cameras and direct observation by staff to track the animals, but that approach left big gaps, and even caused occasional errors in the data. What’s more, the Zoo had to manage unwieldy spreadsheets in the past, to understand their animals’ movements. Since their software could handle only 15 days’ worth of data at a time, that made insights from long-term data (such as behavioral changes in an elephant as it aged) near impossible. The earlier system also made it impossible to integrate observations with external data such as weather changes, fluctuations in the zoo’s attendance, and the like.

The introduction of elephant “ankle bracelets” powered by RFID technology has decidedly changed things for the better. Since elephants are trained to show their feet to handlers for exams and pedicures, getting the bands on them is simple, quick, and stress-free. The Zoo is now able to track a very wide range of parameters for each animal. Nancy Scott, the Coordinator of Elephant Behavioral Science at the Dallas Zoo, now knows that her elephants each walk an impressive average of 10 miles a day. She also knows that Congo, whom she’s dubbed “The Great Explorer”, can walk nearly 17 miles a day, and is also the first one out of the gate when given access to an adjacent habitat to mingle with other species such as giraffe, zebra, and ostrich. That’s useful information to measure the health of elephants not only against their own histories but also against the typical range of the herd. Scott now knows where in the five-acre Giants of the Savanna exhibit the elephants like to go, and where they don’t. She knows who’s been frequenting the mud wallows, pools, scratching posts, log piles or shady spots. She also has a keener understanding of whether the elephants have enough space and how they are using that space, so she can help devise ways to optimize their use of the exhibit.

Elephants being highly social, this technology is also helping the zoo better understand their interactions. They can see which elephants are loners (keeping their distance) vs. which ones are potential friends (frequently traveling together or stationary at night together). When an elephant suddenly moves more slowly or stays near one spot, Scott knows the animal may be ill, leading to faster diagnoses and better health outcomes.


US Medical IT, a Microsoft solution provider and part of the startup community at the University of Texas at Dallas (UTD) Venture Development Department, was an important partner in this effort. With the UTD Venture Development Center’s financial support and US Medical IT’s expertise, the Dallas Zoo enhanced its RFID system with key components of the Microsoft cloud. A SQL Server 2016 -based data warehouse hosted on Microsoft Azure synchronizes the RFID data daily and links it to five other data sources. The data is then made available to Power BI for analysis and to other reporting services running in Azure. The results of the analyses are displayed on dashboards on PCs and mobile devices, including on Apple watches, making insights available to handlers working in the exhibits, to visitors using proposed information kiosks, and to Scott, no matter where she happens to be.

The Zoo can now collect and analyze data across multiple years rather than just days. Information for additional internal and external data sources such as weather, zoo attendance, moon cycles and more can be factored into the analyses. And the best part is that staffers avoid the need for the setup and maintenance of computer systems. The success of the solution has already led Scott to consider ways to expand it further. The addition of Azure Machine Learning, for example, can enable the Zoo to anticipate their elephants’ future needs. The technology can be expanded to other animals, including giraffes, ostriches, and zebras. Gorillas and other apes and monkeys pose an interesting challenge to the technology, because they also move in a third dimension, when they climb trees or other structures. Scott is interested in exploring how their solution could be enhanced to take that into account. You can learn more about the Dallas Zoo solution here.

Institutions around the world involved in the area of animal care are getting inspired by the Dallas Zoo’s pioneering work.

Financial Fabric Helps Hedge Funds Leverage Big Data Analytics Securely, in the Cloud

Since the financial crisis of 2008, the cost of managing hedge funds has grown in pace with the increased regulatory requirements that funds are now expected to meet. According to Preqin, a data provider for the alternative assets industry, manually gathering data from disparate silos, analyzing information, and creating reports can eat up more than 70 percent of a small or midsize hedge fund’s operating budget. On top of that, any discrepancies arising from manual data-handling processes and disjointed workflows can leave a fund vulnerable to regulatory penalties and ultimately to the loss of business.

Hedge funds used to handle many of their IT responsibilities in-house in the past, but it is increasingly clear to modern fund operators that they need much more robust IT infrastructure, including data platforms with advanced analytics capabilities and strong cybersecurity protection, and best practices for security, regulatory compliance and portfolio management.

Financial Fabric is a company that offers hedge funds, institutional investors, and other financial organizations a centralized way to store, analyze and report investment data, using cloud services. To meet the needs of their customers, Financial Fabric required a technology platform that enables fund managers to make data-driven investment decisions without compromising security and privacy. The company decided to base their DataHub solution on Microsoft Azure, taking advantage of its many security features, including the ‘Always Encrypted’ capability of Azure SQL Database.

DataHub includes a client-dedicated data warehouse that ingests information from multiple service providers and systems including prime brokers, fund administrators, order management systems, and industry data sources. In the past, analysts typically downloaded files and documents manually in various formats, and then painstakingly gathered the information into spreadsheets and other tools. Instead, in the new solution, information from diverse sources is automatically collected, cleansed, normalized and loaded to the DataHub, providing up-to-date and accurate information in one place.

Encrypted and stored in the cloud, the information is continuously available to a hedge fund’s business users – including portfolio managers, risk managers, analysts, chief operations officers, and chief financial officers – through business intelligence tools connected to Microsoft SQL Server Analysis Services. Working with interactive Microsoft Power BI dashboards in Excel workbooks, Financial Fabric’s data science team can securely collaborate with clients and create analytics and reports. The DataHub also enables clients to quickly and easily create custom analytics and reports themselves, without IT help. They can also automate workflows such as reconciling data across trades, investment holdings or positions, and cash and margins.

Accessible from virtually anywhere, the analytics and reports are hosted on a Microsoft SharePoint Server farm running on Azure VMs. Historically, the ability to share information on demand while keeping it secure has been an elusive goal for investment managers, but, with DataHub, they are able to securely share information with clients and data scientists to build a more data-driven business with significantly lowered risk. Financial Fabric uses Azure Active Directory and Azure Multi-Factor Authentication to control access throughout all layers of DataHub. The built-in security capabilities have played a critical role in boosting the financial sector’s confidence in cloud solutions.

finfabric-2

Financial Fabric’s primary goal was to solve a business challenge, not deal with technical issues. With Azure, the company and its clients can get on the fast track to data science and avoid spending months and millions of dollars buying or creating software. “We have more data scientists on our team than software developers,” says Subhra Bose, CEO at Financial Fabric. “And they’re focused on the clients’ data, calculating things like investment performance and risk exposure. We have also completely separated our platform development from the data analytics on Azure. That’s given us a tremendous amount of mileage, because we can onboard a client without writing a single line of code.”

One of Financial Fabric’s customers, Rotation Capital, chose to bypass the traditional application-specific approach to developing an institutional infrastructure in favor of DataHub. Within a month, the firm gained a powerful, highly secure data platform with minimal investment in IT staff, software, servers, and other operational overhead. Biagio Iellimo, Controller at Rotation Capital, notes, “Software implementation in the hedge fund industry is a huge pain point. Implementations traditionally take anywhere from six months to a year and a half. So the fact that we were up and running on the Financial Fabric DataHub platform within four weeks is beyond impressive.”

DataHub is a cost-effective, scalable and flexible solution that’s helping hedge funds like Rotation Capital take advantage of big data analytics and protect confidential information while meeting ever-changing business requirements. You can read more about the Financial Fabric solution here.

The availability of secure, cloud-based analytics is proving to be fundamentally transformative for the financial services industry.

CIML Blog Team

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2017/01/11/how-adobe-dallas-zoo-and-financial-fabric-are-using-microsoft-cognitive-services-and-the-azure-data-platform/

How Adobe, Dallas Zoo and Financial Fabric Are Using Microsoft Cognitive Services and the Azure Data Platform

A montage of recent customer use cases, demonstrating the range of capabilities of Microsoft’s AI, Big Data & Advanced Analytics offerings.

Adobe Helps Customers Break the Language Barrier

Adobe Experience Manager (AEM) is an enterprise content management platform that helps large and mid-size companies manage their web content, digital assets, online communities and more. Digital marketers use AEM to manage and personalize online content, including user-generated content on community sites. AEM helps companies build their brands online, driving demand and expanding their markets.

Companies using AEM often serve customers in many countries and languages, and customers were looking to Adobe for an easy way for their digital marketers to translate multinational web sites and content from international user communities into other languages. The sheer volume of information being posted online and prohibitive costs of human translation services meant that most companies could not afford the time and budget needed for manual translation efforts.

adobe-experience-manager

Adobe decided to address the problem by integrating AEM with Microsoft Translator. The Microsoft research team behind the Translator technology had spent over a decade to build a linguistically informed, statistical machine translation service that learns from prior translation efforts. What’s more, the service was flexible, reliable and performant at a massive scale, having performed billions of translations daily, and all while respecting customer data privacy.

To make the system work, all that Adobe’s customers had to do was supply the system with a set of parallel documents, containing the same information in two languages (i.e. the source and target languages). The system would then analyze this material and build a statistical model layered on top of Translator’s generic language knowledge. Translator’s statistical models and algorithms allow the system to automatically detect correlations between source and target language in the training data, helping it determine the best translation of a new input sentence.


AEM versions 6.0 and above now ship with a pre-configured Translator connector, and include a free trial license of 2 million characters per month, enabling users to start automatic translation with minimal effort. Adobe customers can now build, train, and deploy customized translation systems that understand a preferred terminology or style that is specific to their industry or domain. Some customers are also using this new capability to further expand their audience, providing content in previously underserved languages where there is a latent demand or audience. You can learn more about the Adobe solution here.

As people all over the world increasingly look to online communities for instant answers and information, they do not want to be restricted by their language skills. By partnering with Microsoft, Adobe is helping customers become productive by breaking through language barriers.

Dallas Zoo Tracks Elephant Behavior with Fitness Bands

Zoos worldwide have been working hard to provide larger, more natural and varied environments for their elephants. The Dallas Zoo, however, decided to go an extra step further: A growing number of elephants at the Dallas Zoo are now part of a pioneering application that uses RFID bracelets and Microsoft Azure to better understand elephant behavior and provide more customized care.

The earlier solution deployed by the Zoo relied on a combination of video cameras and direct observation by staff to track the animals, but that approach left big gaps, and even caused occasional errors in the data. What’s more, the Zoo had to manage unwieldy spreadsheets in the past, to understand their animals’ movements. Since their software could handle only 15 days’ worth of data at a time, that made insights from long-term data (such as behavioral changes in an elephant as it aged) near impossible. The earlier system also made it impossible to integrate observations with external data such as weather changes, fluctuations in the zoo’s attendance, and the like.

The introduction of elephant “ankle bracelets” powered by RFID technology has decidedly changed things for the better. Since elephants are trained to show their feet to handlers for exams and pedicures, getting the bands on them is simple, quick, and stress-free. The Zoo is now able to track a very wide range of parameters for each animal. Nancy Scott, the Coordinator of Elephant Behavioral Science at the Dallas Zoo, now knows that her elephants each walk an impressive average of 10 miles a day. She also knows that Congo, whom she’s dubbed “The Great Explorer”, can walk nearly 17 miles a day, and is also the first one out of the gate when given access to an adjacent habitat to mingle with other species such as giraffe, zebra, and ostrich. That’s useful information to measure the health of elephants not only against their own histories but also against the typical range of the herd. Scott now knows where in the five-acre Giants of the Savanna exhibit the elephants like to go, and where they don’t. She knows who’s been frequenting the mud wallows, pools, scratching posts, log piles or shady spots. She also has a keener understanding of whether the elephants have enough space and how they are using that space, so she can help devise ways to optimize their use of the exhibit.

Elephants being highly social, this technology is also helping the zoo better understand their interactions. They can see which elephants are loners (keeping their distance) vs. which ones are potential friends (frequently traveling together or stationary at night together). When an elephant suddenly moves more slowly or stays near one spot, Scott knows the animal may be ill, leading to faster diagnoses and better health outcomes.


US Medical IT, a Microsoft solution provider and part of the startup community at the University of Texas at Dallas (UTD) Venture Development Department, was an important partner in this effort. With the UTD Venture Development Center’s financial support and US Medical IT’s expertise, the Dallas Zoo enhanced its RFID system with key components of the Microsoft cloud. A SQL Server 2016 -based data warehouse hosted on Microsoft Azure synchronizes the RFID data daily and links it to five other data sources. The data is then made available to Power BI for analysis and to other reporting services running in Azure. The results of the analyses are displayed on dashboards on PCs and mobile devices, including on Apple watches, making insights available to handlers working in the exhibits, to visitors using proposed information kiosks, and to Scott, no matter where she happens to be.

The Zoo can now collect and analyze data across multiple years rather than just days. Information for additional internal and external data sources such as weather, zoo attendance, moon cycles and more can be factored into the analyses. And the best part is that staffers avoid the need for the setup and maintenance of computer systems. The success of the solution has already led Scott to consider ways to expand it further. The addition of Azure Machine Learning, for example, can enable the Zoo to anticipate their elephants’ future needs. The technology can be expanded to other animals, including giraffes, ostriches, and zebras. Gorillas and other apes and monkeys pose an interesting challenge to the technology, because they also move in a third dimension, when they climb trees or other structures. Scott is interested in exploring how their solution could be enhanced to take that into account. You can learn more about the Dallas Zoo solution here.

Institutions around the world involved in the area of animal care are getting inspired by the Dallas Zoo’s pioneering work.

Financial Fabric Helps Hedge Funds Leverage Big Data Analytics Securely, in the Cloud

Since the financial crisis of 2008, the cost of managing hedge funds has grown in pace with the increased regulatory requirements that funds are now expected to meet. According to Preqin, a data provider for the alternative assets industry, manually gathering data from disparate silos, analyzing information, and creating reports can eat up more than 70 percent of a small or midsize hedge fund’s operating budget. On top of that, any discrepancies arising from manual data-handling processes and disjointed workflows can leave a fund vulnerable to regulatory penalties and ultimately to the loss of business.

Hedge funds used to handle many of their IT responsibilities in-house in the past, but it is increasingly clear to modern fund operators that they need much more robust IT infrastructure, including data platforms with advanced analytics capabilities and strong cybersecurity protection, and best practices for security, regulatory compliance and portfolio management.

Financial Fabric is a company that offers hedge funds, institutional investors, and other financial organizations a centralized way to store, analyze and report investment data, using cloud services. To meet the needs of their customers, Financial Fabric required a technology platform that enables fund managers to make data-driven investment decisions without compromising security and privacy. The company decided to base their DataHub solution on Microsoft Azure, taking advantage of its many security features, including the ‘Always Encrypted’ capability of Azure SQL Database.

DataHub includes a client-dedicated data warehouse that ingests information from multiple service providers and systems including prime brokers, fund administrators, order management systems, and industry data sources. In the past, analysts typically downloaded files and documents manually in various formats, and then painstakingly gathered the information into spreadsheets and other tools. Instead, in the new solution, information from diverse sources is automatically collected, cleansed, normalized and loaded to the DataHub, providing up-to-date and accurate information in one place.

Encrypted and stored in the cloud, the information is continuously available to a hedge fund’s business users – including portfolio managers, risk managers, analysts, chief operations officers, and chief financial officers – through business intelligence tools connected to Microsoft SQL Server Analysis Services. Working with interactive Microsoft Power BI dashboards in Excel workbooks, Financial Fabric’s data science team can securely collaborate with clients and create analytics and reports. The DataHub also enables clients to quickly and easily create custom analytics and reports themselves, without IT help. They can also automate workflows such as reconciling data across trades, investment holdings or positions, and cash and margins.

Accessible from virtually anywhere, the analytics and reports are hosted on a Microsoft SharePoint Server farm running on Azure VMs. Historically, the ability to share information on demand while keeping it secure has been an elusive goal for investment managers, but, with DataHub, they are able to securely share information with clients and data scientists to build a more data-driven business with significantly lowered risk. Financial Fabric uses Azure Active Directory and Azure Multi-Factor Authentication to control access throughout all layers of DataHub. The built-in security capabilities have played a critical role in boosting the financial sector’s confidence in cloud solutions.

finfabric-2

Financial Fabric’s primary goal was to solve a business challenge, not deal with technical issues. With Azure, the company and its clients can get on the fast track to data science and avoid spending months and millions of dollars buying or creating software. “We have more data scientists on our team than software developers,” says Subhra Bose, CEO at Financial Fabric. “And they’re focused on the clients’ data, calculating things like investment performance and risk exposure. We have also completely separated our platform development from the data analytics on Azure. That’s given us a tremendous amount of mileage, because we can onboard a client without writing a single line of code.”

One of Financial Fabric’s customers, Rotation Capital, chose to bypass the traditional application-specific approach to developing an institutional infrastructure in favor of DataHub. Within a month, the firm gained a powerful, highly secure data platform with minimal investment in IT staff, software, servers, and other operational overhead. Biagio Iellimo, Controller at Rotation Capital, notes, “Software implementation in the hedge fund industry is a huge pain point. Implementations traditionally take anywhere from six months to a year and a half. So the fact that we were up and running on the Financial Fabric DataHub platform within four weeks is beyond impressive.”

DataHub is a cost-effective, scalable and flexible solution that’s helping hedge funds like Rotation Capital take advantage of big data analytics and protect confidential information while meeting ever-changing business requirements. You can read more about the Financial Fabric solution here.

The availability of secure, cloud-based analytics is proving to be fundamentally transformative for the financial services industry.

CIML Blog Team

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2017/01/11/how-adobe-dallas-zoo-and-financial-fabric-are-using-microsoft-cognitive-services-and-the-azure-data-platform/

Add Intelligence to Any SQL App, with the Power of Deep Learning

Re-posted from the SQL Server blog.

Recent results and applications involving Deep Learning have proven to be incredibly promising, and across a diverse set of areas too, including speech recognition, language understanding, computer vision and more. Deep Learning is changing customer expectations and experiences around a variety of products and mobile apps, whether we’re aware of it or not. That’s definitely true of Microsoft apps you’re likely to be using every day, such as Skype, Office 365, Cortana or Bing. As we’ve mentioned before, our Deep Learning based language translation in Skype was recently named one of the 7 greatest software innovations of 2016 by Popular Science, a true technological milestone, with machines now sitting at or above human parity, when it comes to recognizing conversational speech.

As a result of these developments, it’s only a matter of time before intelligence powered by Deep Learning becomes an expectation of any app.

In a new blog post, Rimma Nehme addresses the question of how easy might it be for your typical SQL Server developer to integrate Deep Learning into their app. This question is especially timely in light of the recent enhancement to SQL Server 2016 through the integration of R Services, with powerful ML functions, including deep neural networks (DNNs) as a core part of it.

Can we help you turn any SQL app into a truly ‘intelligent’ app, and ideally with just a few lines of code?

To find out, read the original blog post here – the answer may surprise you.


CIML Blog Team

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2017/01/06/add-intelligence-to-any-sql-app-with-the-power-of-deep-learning/

Moving eBird to the Azure Cloud

Re-posted from the Azure Data Lake & HDInsight blog.

Hosted by the Cornell Lab of Ornithology, eBird is a citizen science project that allows birders to submit observations to a central database. Birders seek to identify and record the birds that they discover, and can also report how much effort it took to find those birds. eBird’s web and mobile apps make data recording and interaction super convenient. eBird has accumulated over 350 million records of birds all over the world in the past 14 years.

What’s more, birds are strong indicators of environmental health. They use a variety of habitats, respond to seasonal and environmental cues in specific ways, and undergo dramatic migrations across the globe. Understanding their distribution, abundance and movements across large geographic areas over long periods of time, researchers can build models to understand these patterns, monitor trends and identify conservation priorities.

 Species distribution model showing an abundance of tree swallows throughout an entire year. The model was generated using information collected entirely by eBirders. (Image courtesy of eBird and the Cornell Lab of Ornithology.)

Although the eBird project was providing research opportunities at a scale that would have been inconceivable otherwise, it ran into challenges to do with data growth and the time it took to run analytics models. The project, which has thus far captured 25 million hours of bird observation, faced exponential growth in data volumes. The mid-sized high performance computers being used to run these analytics models were taking as many as 3 weeks to process the results for a single species. That made it very inefficient to generate the results that the researchers needed for the 700 odd species of birds that regularly inhabit North America.

Thanks to a recent collaboration between the Cornell Lab and Microsoft, this project and the associated machine learning workflow were migrated to the fully managed, highly scalable Azure HDInsight (Hadoop) service, a key component of the Microsoft Cortana Intelligence Suite. As a result of this partnership, researchers were able to scale their clusters sufficiently to reduce analysis run times to as little as 3 hours, generating results across more species dramatically faster. This, in turn, provides more timely results for conservation staff to then use in their planning process. They have also been able to run models on dozens more species than they would have otherwise.

The complete solution is built on Azure Storage, HDInsight, Microsoft R Server, Linux Ubuntu, Apache Hadoop MapReduce and Spark.


You can click on this link to read the original post, or on the architecture diagram above. 

Taking advantage of the scalability, manageability and open source support of the Microsoft Azure cloud platform, the researchers behind eBird hope to drive further innovation and accelerate their research and conversation efforts, working closely with the community. 

CIML Blog Team

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2017/01/05/moving-ebird-to-the-azure-cloud/

Hello 2017, and Recap of Top 10 Posts of 2016

As we kick off what will surely be another very exciting year of progress in artificial intelligence, machine learning and data science, we start with a quick recap of our “Top 10” most popular posts (based on aggregate readership) from the year just concluded.

Here are the posts that had the most page views in the recently concluded year, in increasing order of readership:

10. Recent Updates to the Microsoft Data Science Virtual Machine

We announced a major set of updates to the Data Science VM (DSVM) in September last year. DSVM gives you a comprehensive set of tools for data movement, storage, exploration/visualization, modeling with ML/AI algorithms, and operationalization – and using multiple languages in either Linux or Windows environments.


9. Introducing Microsoft R Server 9.0

Just last month, we announced our latest and most powerful version of Microsoft R Server. Supporting popular operating systems and a variety of data sources, MRS 9.0 helps you create and deploy sophisticated analytics models for real world problems, efficiently and at scale.


8. Microsoft and Liebherr Collaborating on New Generation of Smart Refrigerators

An unexpected post for a Top 10 list, perhaps, but this goes to show the broad excitement in our community around the possibilities of computer vision, deep learning and IoT. Our award-winning vision and deep learning capabilities around image processing are now a cornerstone of an ever-widening array of products offered both by Microsoft directly and our customers and partners.


7. Introducing the Team Data Science Process from Microsoft

In October, we introduced the Team Data Science Process – a methodology and set of practices designed to help your business truly reap the benefits promised by collaborative data science.


6. Game On! Introducing Cortana Intelligence Competitions

Back in March last year, we gamified Cortana Intelligence with the launch of our competition platform, and announced our first ever contest, a very successful competition on Decoding Brain Signals.


5. A One-Step Program for Becoming a Data Scientist

A link to a webinar we delivered early last year, in response to an ask from the community on what it takes to be a successful data scientist. Both the webinar and blog post emphasize the importance of building things first-hand.

4. Building Deep Neural Networks in the Cloud with Azure GPU VMs, MXNet and Microsoft R Server

This was the first in a series of posts we launched in September last year, showcasing deep learning workflows on Azure. In this post, we setup N-Series VMs on Azure with NVIDIA CUDA and cuDNN support, using MXNet as one example of deep learning frameworks that can run on Azure. We also show how Microsoft R Server can harness the deep learning capabilities of MXNet and Azure GPUs using simple R scripts.

3. Data Science with Microsoft SQL Server 2016 – Free eBook

Few things in life can beat “free”, and that was certainly true about our free eBook on creating intelligent apps using SQL Server and R. You can now embed intelligent analytics and data transformations right in your database, and make transactions intelligent in real time. Combining the performance of SQL Server in-memory OLTP and in-memory columnstores with R and machine learning, apps can achieve extraordinary analytical performance in production – and with all the benefits you expect from our industrial-strength database, including high throughput, parallelism, security, reliability, compliance certifications and great manageability.


2. Announcing R Tools for Visual Studio

The ML and data science community was just as delighted as us when we announced the newest language spoken by Visual Studio, back in March 2016.


1. Making R the Enterprise Standard for Cross-Platform Analytics, Both On-Premises and in the Cloud

Early in 2016, we announced how we are delivering Microsoft R Server across multiple platforms, allowing enterprise customers to standardize advanced analytics on one core tool, and regardless of whether they are using Hadoop, Linux or Teradata. We also announced that, on Windows, Microsoft R Server (MRS) would be included in SQL Server 2016. This post also talked about the free Developer Edition of MRS, as well as our commitment to Microsoft R Open. This was our #1 most widely read and circulated post of 2016.

Stay tuned to this channel for much more exciting news in 2017.

We wish all our readers a very happy and prosperous new year. 

CIML Blog Team

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2017/01/03/hello-2017-and-recap-of-top-10-posts-of-2016/

Singapore Machine Learning & Data Science Summit – Recap

This post is authored by Tamarai G V, Senior Product Marketing Manager at Microsoft.

Singapore has started to embrace the many benefits of digital transformation, and data plays a central role in this process. From using non-traditional indicators such as electricity consumption and public transportation to monitor the economy to helping the government improve the lives of ordinary citizens, machine learning and data science are being put to use to solve real world problems.

As part of Singapore’s digital efforts, and to nurture a vibrant ML and data science community, the inaugural Machine Learning and Data Science Summit in Asia was held in Singapore on Dec 9th, 2016, at the beautiful University Town in the National University of Singapore (NUS).

The event was jointly organized by the Government Technology Agency of Singapore (GovTech), NUS and Microsoft, and attended by hundreds of data scientists, developers, students and faculty. The full day summit had an exciting agenda, with keynotes, breakout sessions and hands-on labs helping attendees learn how to tap the power of Cortana Intelligence, Microsoft R Server and SQL Server R Services to build intelligent applications. There were sessions on building intelligent bots, demystifying deep learning, understanding how the Team Data Science Process can help jump-start successful data science teams and many more.


The summit kicked off with keynote sessions by Jessica Tan, Managing Director for Microsoft Singapore, Chan Cheow Hoe, CIO at GovTech, and Professor Lakshminarayanan of NUS. Jessica highlighted the new possibilities of digital transformation and the need to approach data sciences as a team sport.


Chan Cheow Hoe spoke about the use of Data Science in the Public Sector, and how a data-driven approach has solved real world problems in Singapore, such as the recent Circle Line incidents. He also shared how data science can be harnessed to derive deep insights from data to inform policy changes and reviews, and to improve operations and service delivery through applications and data visualisation.


Finally, Professor Lakshminarayanan welcomed the attendees to University Town, and shared the work that college is doing on systems thinking and design, and how it is relevant to the data science community.


Hongyi Li, Product & Engineering lead at GovTech, presented on how his organisation is working on using data for the public good and how open data can help citizens understand and use data. He talked about how data.gov.sg wants to help agencies establish common data sharing infrastructure and make it accessible to use for decision making.


Other sessions that followed included topics on adopting a system thinking towards data science by Wee Hyong Tok and Jenson Goh (from NUS); Matt Winkler and Jennifer Marsman who shared how one can bring intelligence into applications using Cognitive Services and Cortana Intelligence Suite; and Anusua Trivedi who demystified deep learning and shared the exciting applications that can be built in this area.

A key highlight of the event was the Hackathon, led by Hang Zhang, where participants from across academia and industry pitted their skills against the best in ML and data science. Hackathon participants tackled the problem of predicting the number of fatalities in traffic accidents in which drunk drivers had an impact on the outcome. Using the Cortana Intelligence Competition Platform, participants came up with many creative ways of building and improving on their solutions, and worked away to get on the leaderboard.


The Summit concluded with a closing address by Vijay Narayanan, Director of Data Sciences at Microsoft, and an awards ceremony recognizing the hackathon winners.


It was great to see the winners being offered internship opportunities by different organizations at the conclusion of the event!

We look forward to the next Data Science Summit in 2017.

Tamarai

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2016/12/16/singapore-machine-learning-data-science-summit-recap/

Using Cortana Intelligence in HoloLens Applications

This post is authored by Scott Haynie, Senior Software Engineer, and Senja Filipi, Software Engineer, at Microsoft.

Telemetry plays an important role when you operationalize new experiences/apps that span the web, mobile and IoT, including new gadgets such as the Microsoft HoloLens. The abundance of data that is made available can help developers monitor and track system health and usage patterns, and provide important new insights into how users interact with your application. Tapping into this wealth of information can really help you align your customers’ experiences with their needs and expectations.

At the Ignite 2016 Innovation Keynote, we showed the future of home improvement as envisioned by Lowe’s and Microsoft. As part of this experience, we showed how to use telemetry emitted by HoloLens to understand the [virtual] places that customers are gazing at, as they experience an immersive home remodeling experience in 3D, using the augmented reality capabilities of HoloLens.

Kitchen remodeling experience with HoloLens.

Heat map based on HoloLens telemetry data.

In this post, we show how you can use the Cortana Intelligence Suite to ingest data from the HoloLens application, analyze it in real-time using Azure Stream Analytics, and visualize it with Power BI.

Empowering HoloLens Developers with Cortana Intelligence

In HoloLens applications, users can interact through one of these methods:

  1. By gazing at an object.
  2. By using hand gestures.
  3. By using voice commands.

When wearing, and interacting with, the HoloLens, a frontal camera tracks head movements, including the “head ray” and focus. The point that is being looked at is reflected using a cursor as a visual cue. These interactions are handled by the Update method (part of the MonoBehavior class). This method gets called at the refresh rate frequency, usually 60 frames per second. It’s crucial to not slow down the update process with any side operations.

We set out with a simple goal that HoloLens developers should be able to use the Cortana Intelligence Suite and any related Azure services with just a few lines of code. In this example, you can see the code in one of the Update methods that tracks the gaze of the HoloLens user. To use Cortana Intelligence, the HoloLens application needs to just add this one line:

logger.AddGazeEvent(FocusedObject.gameObject.name)

The HoloLens is now able to send telemetry data to Event Hub, and you can further analyze the data using various Azure Services, for instance, Azure Stream Analytics.


Calling the telemetry AddGazeEvent from the HoloLens app.

Building the End to End Solution

A canonical telemetry processing pipeline consists of the following:

  1. An Event Hub that enables the ingestion of data from the HoloLens client application.
  2. A Stream Analytics job that consumes the telemetry data, analyzes it real time, and writes the insights derived into Power BI, as an output.
  3. A PowerBI dashboard.

If you don’t have an Azure account yet, you can obtain a free account which will help you follow along in this post. Azure provides a rich set of libraries that developers can use to interact its many services. When using these libraries with a HoloLens application (which is a Universal Windows App), you will not be able to use NuGet packages directly if they rely on the .NET full framework implementation. Additionally, Azure services provide REST interfaces for client communication.

In our case, to send telemetry data from the HoloLens to the Event Hub, we implemented a .NET library, using the core .NET framework to meet the UWP apps requirements. The DLL handles the batching of events and composing of the request payload, and takes care of network retries.


Initializing the Telemetry Library

Step-by-step instructions on how to set up the Azure services and send data to the Azure Event Hub with retries from the UWP app can be found in GitHub here. Additional resources that you may find useful are included below. We would love to hear from you, so do let us know what you think – you can send your feedback via the comments section below.

Scott & Senja

Resources:

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2016/12/16/using-cortana-intelligence-in-hololens-applications/

Exploring Azure Data with Apache Drill, Now Part of the Microsoft Data Science Virtual Machine

This post is authored by Gopi Kumar, Principal Program Manager in Microsoft’s Data Group.

We recently came across Apache Drill, a very interesting data analytics tool. The introduction page to Drill describes it well:

“Drill is an Apache open-source SQL query engine for Big Data exploration. Drill is designed from the ground up to support high-performance analysis on the semi-structured and rapidly evolving data coming from modern Big Data applications, while still providing the familiarity and ecosystem of ANSI SQL, the industry-standard query language.”.

Drill supports several data sources ranging from flat files, RDBMS, NoSQL databases, Hadoop/Hive stored on local server/desktop or cloud platforms like Azure and AWS. It supports querying various formats like CSV/TSV, JSON, relational tables, etc. all from the familiar ANSI SQL language (SQL remains one of the most popular languages used in data science and analytics). The best part of querying data with Drill is that the data stays in the original source and you can join data across multiple sources. Drill is designed for low latency and high throughput, and can scale from a single machine to thousands of nodes.

We are excited to announce that Apache Drill is now pre-installed on the Data Science Virtual Machine (DSVM). The DSVM is Microsoft’s custom virtual machine image on Azure, pre-installed and configured with a host of popular tools that are commonly used in data science, machine learning and AI. Think of DSVM as an analytics desktop in the cloud, serving both beginners as well as advanced data scientists, analysts and engineers.

Azure already provides several data services to store and process analytical data ranging from blobs, files, relational databases, NoSQL databases, and Big Data technologies supporting varied types of data, scaling / performance needs and price points. We wanted to demonstrate how easy it is to setup Drill to explore data stored on four different Azure data services – Azure Blob Storage, Azure SQL Data Warehouse, Azure DocumentDB (a managed NoSQL database) and Azure HDInsight (i.e. managed Hadoop) Hive tables.

Towards that end, we’ve published a tutorial on the Cortana Intelligence Gallery that walks you through the installation and how to query data with Drill. the tutorial that will guide you through the steps to set up connections from Drill to different Azure Data services.

Drill also provides an ODBC/JDBC interface, allowing you to perform data exploration on your favorite BI tool such as Excel, Power BI or Tableau, using SQL queries. You can also query data from any programming language such as R or Python with ODBC/JDBC interfaces.

While on the Data Science Virtual Machine, we encourage you to also take a look at other useful tools and samples that come pre-built. If you’re new to the DSVM (which is available in Windows and Linux editions, plus a deep learning extension to run on Nvidia GPUs), we invite you to give the DSVM a try through a Azure free trial. We also have a timed test drive, available for the Linux DSVM now, that does not require an Azure account. You will find more resources to get you started with the DSVM below.

In summary, Apache Drill can be a powerful tool in your arsenal, and can help you be nimbler with your data science projects and gain faster business insights on your big data. Data scientists and analysts can now start exploring data in its native store without having to wait for ETL pipelines to be built, and without having to do extensive data prep or client side coding to bring together data from multiple sources. This can be a huge boost to your teams’ agility and productivity.

Gopi

Windows Edition:

Linux Edition:

Webinar:

https://channel9.msdn.com/blogs/Cloud-and-Enterprise-Premium/Inside-the-Data-Science-Virtual-Machine (Duration: 1 Hour)

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2016/12/14/exploring-azure-data-with-apache-drill-now-part-of-the-microsoft-data-science-virtual-machine/

Building Intelligent Bots for Business

This post is authored by Herain Oberoi, Senior Director of Product Marketing at Microsoft.

Earlier today, in San Francisco, we provided an update on how Microsoft is helping to democratize Artificial Intelligence (AI) by making it accessible to everyone and every organization. Today’s focus was on conversational computing, which combines the power of natural language with advanced machine intelligence to help people engage with technology in more natural and personal ways.

As we talk to businesses and governments who are looking to take advantage of these new capabilities, we see significant value being created when leading organizations start using intelligent bots to transform business processes such as customer services, helpdesks, and even factory floor operations.

One example is at Rockwell Automation, which provides industrial automation and information solutions to customers in more than 80 countries. Their customers wanted access to information in their production lines in faster and more innovative ways, and so, with that objective in mind, Rockwell Automation used the Bot Framework and Cognitive Services in Cortana Intelligence to build Shelby™, a bot that monitors production more efficiently and lets managers know the status of their operations through more natural forms of interaction.

“Our customers need to move quickly to meet their goals. Shelby™ gives them an entirely new way to interact with their environment. The health and diagnostics of their production is critical to make the decisions that matter.”

Paula Puess, Global Market Development Manager, Rockwell Automation

Australia’s Department of Human Services (DHS), an arm of the government responsible for delivering social and health related services and payments, is pioneering a proof of concept to deliver intelligent customer experiences powered by deep learning. Using the machine learning and cognitive services capabilities in Cortana Intelligence, DHS is building an ‘expert system’ that helps its employees respond faster and more effectively to citizen queries by infusing bots with deeper human context and conversational understanding, ultimately improving and expanding their customer engagement channels.

Improving Customer Interactions with Intelligent Bots

Intelligent bots deepen customer engagement by augmenting the skills and knowledge of employees interacting with customers, and also via direct conversations that provide for more natural and personalized interactions at massive scale. Specifically:

  1. Bots go beyond simple task completion, using social and historical context to better infer intent and make recommendations that are actionable in the context of the conversation.
  2. Bots drive efficiencies by automating workflows and integrating task completion with existing systems in the context of the business process.
  3. Bots help uncover new insights about customer challenges and preferences by being able to capture and reason over all the customer interaction data.

The illustration below shows how businesses can leverage intelligent bots to augment their contact center operations and improve efficiencies.

bots-h

The use of an intelligent bot in this example helps the business in the following ways:

  • A majority of common customer requests can be fulfilled by bots. Customer requests that require deeper human intervention are filtered and handed off to contact center agents. This intelligent filtering helps reduce costs while simultaneously improving the customer service experience.
  • The quality of contact center agent interactions is improved as bots help augment the skills and knowledge of agents by providing real-time recommendations in the context of the current conversation.
  • The business also benefits from enhanced insights obtained by analyzing the rich customer interaction data captured by the bots. This can help the business spot emerging patterns, take preemptive actions on issues, and much more.

Getting Started

Developers can get started with the open source based Microsoft Bot Framework which includes the Bot Builder SDK, Bot Connectors, Developer Portal, Bot Directory and an emulator to use and test your bots.

As you build, test and scale your bots in the cloud, the Microsoft Azure Bot Service helps you accelerate your work through an integrated environment that is purpose-built for bot development. The Azure Bot Service allows you to get started quickly with built-in templates, scale and pay on demand as your needs grow over time, and reach users on multiple channels including from your app or website, via text/SMS, Skype, Slack, Facebook Messenger, Kik, Office 365 email, and other popular services.

Building an intelligent bot goes beyond simple task completion – a complete business solution requires cognitive understanding, integration with business processes, and the ability to gain deep insights on customer interaction data. Microsoft Cortana Intelligence provides you with all the capabilities you need, including big data storage, orchestration, advanced analytics and cognitive services, to build your intelligent bots.

I hope customers like Rockwell Automation and Australia’s Department of Human Services give you the needed inspiration to take the leap and start defining the requirements for intelligent bots that can help improve customer engagement at your organization.

Herain

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2016/12/13/building-intelligent-bots-for-business/

Using SQL Server 2016 with R Services for Campaign Optimization

This post is authored by Nagesh Pabbisetty, Partner Director of Program Management at Microsoft.

We are happy to announce a new Campaign Optimization solution based on R Services in SQL Server 2016, designed to help customers apply machine learning to increase response rates from their leads.

Marketing organizations use a number of ways to reach and interact with customers. In addition to providing offers that are best suited for a given target segment, they also make use of the most appropriate communication medium, such as SMS, email or phone calls. Our new Campaign Optimization solution reviews a set of offers alongside a set of prospective customers and campaign business rules to figure out which offer should go out to which prospect, on what channels, and precisely when.

We have published this solution in the Cortana Intelligence Solutions Gallery. The solution provides a hands on experience by deploying into your Azure subscription. The deployment takes just a few clicks, getting the solution up and running by configuring it on our most popular VM, namely the Microsoft Data Science VM (DSVM) that comes loaded with all the tools that a data scientist will need. The code is also published on GitHub, so if you prefer to run this on your own machine entirely, you can the instructions that are available there.

We use a real world scenario involving the insurance industry. Model predictors include demographic information for the leads or prospects, historical campaign performance, and product-specific details. The model then predicts the probability that each lead in the database will make a purchase on a particular channel, and we predict this for each day of the week and at different times of the day. Recommendations on which channel, day of the week and time of the day to use, when targeting these users, are based on what the model predicts will have the highest probability of converting into a purchase. If the report needs to be updated, this can be done via the Power BI desktop that comes pre-installed in the VM.

Two Experiences, for Two Different Personas

  • The business manager responsible for the campaign needs leads to be targeted using the optimal channels and get the highest possible response rate. The manager uses the Power BI -powered dashboard to get a snapshot view where they can then quickly determine the right channels to use for each lead.

Snap shot

  • The data scientists who are testing and developing solutions can work from the convenience of their R IDE on their client machine, while pushing the compute to the SQL Server machine. They can also use PowerShell scripts or Jupyter Notebooks, in addition to using IDEs such as R Tools for Visual Studio. Completed solutions are deployed to SQL Server 2016 by embedding calls to R in stored procedures. These solutions can then be further automated with SQL Server Integration Services and the SQL Server agent.

We would love to hear your feedback, so do give the solution a try and write to us via email here.

Nagesh

from Cortana Intelligence and Machine Learning Blog https://blogs.technet.microsoft.com/machinelearning/2016/12/07/using-sql-server-2016-with-r-services-for-campaign-optimization/