Posted on Leave a comment

Meeting the tech demand for security-cleared veteran talent

This post was originally published on LinkedIn and written by U.S. Marine Corps Major General (Ret.) Chris Cortez, vice president of Microsoft Military Affairs, on November 15, 2018.

Today the United States government faces an unprecedented range of threats—from both traditional foes to modern menaces in the digital realm. And as our country faces these new challenges, it is also in the midst of a digital transformation of its services and operations. Equipping people on the front lines of national security with modern skills and technological awareness is paramount, and we’re committed as industry partners to helping along that journey.

Yet, the work of modernizing and protecting the US government is not just a technology story, it is a people story. Our ability to safeguard our people and institutions relies on a highly specialized and skilled workforce to build, code, and innovate for the future. What we’re seeing now though is a gap between the demand for capabilities and the supply of this much-needed talent.

Many of us have heard about the technology skills gap. We know that technology jobs are growing at a faster rate than the average for all other occupations, and that the demand greatly outpaces the supply. The Brookings Institution suggested last year that digitization of the US economy will require significant investment in education and training both to broaden the pipeline of talent and also to ensure that underrepresented groups aren’t left behind in the new digital economy.

In government though, there is an even more specific gap—that of “cleared talent”—those with security clearances that can be applied to tackle problems that require access to highly sensitive information. And within this pool of talent, there is a very small population of cleared talent with the background in technology that is required to help our government modernize at levels needed. Despite the need to expand this workforce, we’ve seen the population of cleared talent decrease by 30 percent since 2013, with wait times for security clearance approvals now at their highest level ever. Last year, the US was sitting on a security clearance backlog of more than 700,000 applicants.

New data we obtained in partnership with LinkedIn shows the demand for cleared software engineers is very high—with the Washington, D.C., area as the highest demand in the country—and that only 5 percent of professionals with a security clearance hold a degree in computer science. The top two industries hiring cleared talent are also expected to grow even more in the next year: The defense and space industry by 53 percent, and the information technology industry by 49 percent.

The stunted pipeline for cleared talent is not just an HR problem. It is a matter of national security and a threat to progress and innovation.

One population that is critical to help close this gap is our country’s talented veteran workforce. We know that 200,000 military members transition out of the service every year, and, of those with security clearances, many are active duty military, according to a 2018 ClearanceJobs.com report.

At Microsoft, we pioneered Microsoft Software & Systems Academy (MSSA), which trains transitioning service members and veterans over the course of an 18-week program for careers in the IT industry. Since MSSA started five years ago, nearly 400 companies and organizations have hired our graduates. For us, it’s proof not only that veterans are an incredible talent pool, but that programs like MSSA are essential to addressing critical skills gaps in the tech industry outside the traditional four-year degree. Companies like Microsoft are better for having these phenomenal veterans working for us and helping advance our mission.

That’s why I am excited to announce that Microsoft will launch a new, dedicated MSSA-cleared talent cohort at Washington state’s Joint Base Lewis-McChord this spring. The program will cater to service members and veterans with active clearances who are interested in learning cloud application development, which is in high demand in the technology industry. As with every MSSA cohort, these students will be mentored by current Microsoft employees and upon completion of the course will be guaranteed an interview with Microsoft or one of our hiring partners, including the US government.

National security depends on private and public institutions working together to prevent and address emerging threats. And being a trusted partner for government transformation means that we are just as committed to investing in our own talent pipeline as we are in theirs. Innovative training programs like MSSA, which benefit both veterans and the industry as a whole, are critical.

There is much more that can be done to address both the technology skills gap and the cleared talent gap, but we must seek new and innovative ways to both build our pipeline of workers and support those who are being left out of the digital economy. Empowering active duty service members and veterans to transition to careers in technology is a win-win for helping to sustain a vital population of our workforce while building a critical, diverse pipeline for in-demand talent. By prioritizing IT and technology training for veterans to operate in the private sector, we can transform our collective capabilities to keep our country safe and moving forward.

Visit military.microsoft.com/MSSA to learn more.

Posted on Leave a comment

SMITE and Paladins esports return to Mixer

Mixer has a long-standing partnership with Hi-Rez, starting with the Paladins Console Series announced in October of last year, and then the announcement of the SMITE Pro League last February. Earlier today at Hi-Rez Expo (HRX), we announced a continued partnership with Hi-Rez Studios and Skillshot Media, which includes the expansion of SMITE and Paladins esports exclusively on Mixer!

We’re excited to share that the 2019 season of the SMITE Pro League and SMITE Console League will continue to be exclusive on Mixer, building on the success of both leagues during Season 5. In addition, the other premium league from Hi-Rez, Paladins Premier League, featuring the top esports organizations from around the world, will also be coming exclusively to Mixer in 2019!

Earlier this year, Mixer implemented interactive stats dashboards for SMITE esports broadcasts using MixPlay custom controls. We will continue to introduce innovative Mixer features, such as MixPlay integrations and custom Skill packs, for all future SMITE and Paladins esports broadcasts.

The latest integration, MixPlay Rewards, where you can earn in-game items just by linking your Mixer and Hi-Rez accounts and watching official Hi-Rez broadcasts, will debut at HRX this weekend for both SMITE and Paladins broadcasts. Tune-in to official Hi-Rez Expo broadcasts for SMITE and Paladins on Mixer to earn HRX Loki and Convention Furia!

We’ll be announcing more exciting news about the 2019 season of SMITE and Paladins esports over the next few days  at HRX  – follow @WatchMixer, @SmitePro, and @PaladinsPro on Twitter for all the latest updates!

Posted on Leave a comment

HoloLens experience pays homage to Mont-Saint-Michel in Normandy, France

See the results of a months-long effort to create a HoloLens experience that pays homage to Mont-Saint-Michel, in Normandy, France, in all its forms – as a physical relief map and work of art; as a real place visited by millions of people over the centuries; and as a remarkable digital story of resilience. In this three-part Today in Technology series, they examine how AI and mixed reality can open a new window into French culture by using technology like HoloLens.

Posted on Leave a comment

Introducing the Azure Blockchain Development Kit

“Developers! Developers! Developers!” That phrase is synonymous with Microsoft’s history of democratizing complex technologies and empowering anyone with an idea to build software.

Over four decades, we’ve lowered barriers to development with developer tooling, enterprise integration, DevOps, PaaS, and SaaS. Today, serverless offerings from Functions and Logic Apps to Azure DevOps and IoT Central remove friction for development in the cloud.

This morning, we’re excited to announce the initial release of the Azure Blockchain Development Kit which is built on Microsoft’s serverless technologies and seamlessly integrates blockchain with the best of Microsoft and third-party SaaS.

This kit extends the capabilities of our blockchain developer templates and Azure Blockchain Workbench, which incorporates Azure services for key management, off-chain identity and data, monitoring, and messaging APIs into a reference architecture that can be used to rapidly build blockchain-based applications.

These tools have become the first step for many organizations on their journey to re-invent the way they do business. Apps have been built for everything from democratizing supply chain financing in Nigeria to securing the food supply in the UK, but as patterns emerged across use cases, our teams identified new ways for Microsoft to help developers go farther, faster.

This initial release prioritizes capabilities related to three key themes: connecting interfaces, integrating data and systems, and deploying smart contracts and blockchain networks.

Connect

To deliver end to end blockchain solutions for consortiums, developers need to enable organizations, people, and devices to connect to the blockchain and do it from a heterogenous set of user interfaces.

Take for example an end to end supply chain for a commodity such as cocoa.

  • SMS and voice interfaces enable small hold farmers in Africa to transact and track their goods at the first mile of the supply chain.
  • Internet of Things (IoT) devices deliver sensor data to track the conditions of the goods at different points in their journey to market – tracking the humidity in the containers where the beans are held to the temperature of the end product of ice cream that it is incorporated into.
  • Mobile clients enable logistics providers to accept and transfer responsibility for products on their journey from manufacturer to retail using the compute power that already exists in the pockets of its employees. Mobile devices also have sensors such as GPS and cameras that can add complementary data that can help attest to the what, where, and when of deliveries.
  • Backend Systems and Data in the form of ERP systems such as Dynamics and SAP are used to manage core processes for different participants. These systems also become clients via extension and need to interact with smart contracts to provide and receive attestable data on behalf of an organization.
  • Bots and assistants enable manufacturers and retailers to interact with the supply chain. This includes interacting with smart contracts for orders and provenance using natural language and using attestable data from the blockchain to direct actions taken on behalf of a user.
  • Web clients enable end consumers to query the origin of the product purchased at retail, typically a mix of provenance and story of their journey of their product from “farm to fork”

The Azure Blockchain Development Kit includes samples for all of these scenarios, including inbound and outbound SMS, IVR, IoT Hub and IoT Central, Xamarin mobile client for iOS and Android, Dynamics integration via Common Data Service (CDS), bots and assistants (Cortana, Alexa, Google Assistant) and web UX.

Integrate

Businesses are using blockchain and smart contracts to facilitate multi-party processes. Blockchain also delivers real-time transparency of the states and events of those contracts to appropriate participants.

End to end blockchain solutions require integration with data, software, and media that live “off chain”. External updates and events can trigger actions on smart contracts. Smart contract events and state changes can then trigger actions and data updates to “off chain” systems and data. These external systems and AI will also need the ability to query attestable data from smart contracts to inform action.

Specifically, there are two areas of integration where guidance is most needed:

Documents and Media: Documents and media do not belong on chain, but business processes often involve images, videos, audio, Office documents, CAD files for 3D printers or other file types.

The common pattern is to generate a unique hash of the media and the metadata that describes it. Those hashes are then placed on a public or private chain. If authenticity of a file is ever questioned, the “off chain” files can be re-hashed at a later time and that hash is compared to the “on chain” hash stored on the blockchain. If the hashes match, the document is authentic, but if so much as a pixel of an image or letter in a document is changed, the hashes will not match and this will make obvious that tampering has occurred.

Today we’re releasing a set of Logic Apps that enable the hashing of files and file related metadata. Also included are smart contracts for files and a file registry to store the hashes on chain.

Logic Apps have been created to deliver this functionality for files added to the most popular sources for documents and media, including Azure Storage, OneDrive, One Drive for Business, SharePoint, Box, Adobe Creative Cloud, and FTP.

Documents and Media

Smart Contract Interaction: Getting blockchain off the whiteboard and into production means dealing with the realities of how counterparties interact today. That reality is that Enterprise integration is messy.

Microsoft brings our decades of experience in this area to blockchain. Our work with integrating Enterprise systems began almost two decades ago with the introduction of BizTalk server, and our focus on database integration traces back to our co-development of Open Database Connectivity (ODBC) in the 1990s. All of our experience has been captured and made available in Azure services. This includes 200+ connectors available in Logic Apps and Flow, and the robust capabilities in our data platform.

Smart Contract Interaction

The Blockchain Application Development Kit includes Workbench integration samples in the following areas:

Logic App Connectors for Blockchain

Today, we are also announcing that we will release a set of Logic App and Flow Connectors to extend these samples to ledgers like Ethereum, Corda, Bitcoin, and others.

“At R3, we are committed to ensuring developers can deploy CorDapps quickly, securely and easily. The Azure Blockchain Development Kit will give our enterprise customers tools to integrate with the applications, software, and devices that people use every day like Outlook, Alexa, SMS, and web UX. Blockchain is moving out of the labs and into everyday business applications.”

– Mike Ward, Head of Product Management, R3

The Ethereum blockchain connector is available today and enables users to deploy contracts, call contract actions, read contract state and trigger other Logic Apps based on events from the ledger.

Logic Apps

Deploy

With the mainstreaming of blockchain technology in enterprise software development, organizations are asking for guidance on how to deliver DevOps for smart contracts and blockchain projects.

Common questions include:

  • My business logic and data schema for that logic is reflected in smart contracts. Smart contracts are written in languages I’m less familiar with like Solidity for Ethereum, and Kotlin for Corda, or Go for Hyperledger Fabric.  What tools can I use to develop those in?
  • How do I do unit testing and debugging on smart contracts?
  • Many blockchain scenarios reflect multi-party transactions and business workflows. These workflows include signed transactions from multiple parties happening in specific sequences. How do I think about data for test environments in that context?
  • Smart contracts are deployed to the blockchain, which is immutable. How do I need to think about things such as infrastructure as code, local dev/test, upgrading contracts, etc.?
  • Blockchain is a data technology shared across multiple organizations in a consortium, what are the impacts on source code control, build and release pipelines in a global, multi-party environment?

While there are some nuances to the approach, the good news is that just like other types of solution development, this model can readily be addressed in a DevOps model.

DevOps Model

Today, we’re announcing the release of the whitepaper, “DevOps for Blockchain Smart Contracts.”

“We’re excited to work with Microsoft to create the canonical DevOps experience for blockchain engineers. Our paper, ‘DevOps for Blockchain Smart Contracts’, goes into rigorous detail and provides examples on how to develop blockchain applications with an eye toward CI/CD in consortium environments.”

– Tim Coulter, Founder of Truffle

Complementing the whitepaper is an implementation guide, available through the Azure Blockchain Development Kit, that shows how to implement CI/CD for smart contracts and infrastructure as code using Visual Studio Code, GitHub, Azure DevOps and OSS from Truffle.

A great platform for blockchain application development

The Azure Blockchain Development Kit is the next step in our journey to make developing end to end blockchain applications accessible, fast, and affordable to anyone with an idea. It is built atop our investments in blockchain and connects to the compute, data, messaging, and integration services available in both Azure and the broader Microsoft Cloud to provide a robust palette for a developer to realize their vision.

Logic Apps and Flow deliver a graphical design environment with more than 200 connectors dramatically simplifying the development of end to end blockchain solutions, and Azure Functions enable the rapid integration of custom code.

A serverless approach also reduces costs and management overhead. With no VMs to manage, built-in scalability, and an approachable pricing model the Azure Blockchain Development Kit is within reach of every developer – from enthusiasts to ISVs to enterprises.

Solutions are written using online visual workflow designers and Visual Studio Code, a free download that provides an integrated development environment on Windows, Mac, and Linux.

The resulting applications will run atop a network that has higher rated cloud performance than other large-scale providers and enable federating identities between participants using Azure Active Directory. With Azure, those applications can be deployed to more regions than any other cloud provider and benefit from more certifications.

We look forward to seeing what you’ll build, and we’ll continue to both listen and look for ways to help as we build a decentralized future together.

To learn more about how to use these samples to build and extend blockchain applications, you can find a host of videos on our Channel 9 show Block Talk.

You can also stay up to date with the latest updates from Azure Blockchain by following us on Twitter @MSFTBlockchain.

Posted on Leave a comment

New updates and technologies will help businesses do more with AI

As AI breakthroughs abound, businesses look to score benefits

When Arccos Golf launched its first performance tracking system for golfers, it combined the telemetry from sensors and a smartphone app to give players detailed data and feedback about every shot.

Knowing how far and how accurately they could hit the ball under different conditions helped players uncover weaknesses and improve their game. But there was so much more that could be done.

“We had an ‘a-ha moment’ about providing a virtual caddie for every player. Just like a human caddie, ours would know the player, know the course, know the weather and provide the player with a club recommendation,” said Jack Brown, senior vice president of product & software at Arccos Golf. “So we thought, ‘Why don’t we use AI to create a virtual caddie?’”

Read more

Posted on Leave a comment

Microsoft agrees to acquire conversational AI and bot development company, XOXCO

Conversational AI is quickly becoming a way in which businesses engage with employees and customers: from creating virtual assistants and redesigning customer interactions to using conversational assistants to help employees communicate and work better together. According to Gartner, “By 2020, conversational artificial intelligence will be a supported user experience for more than 50 percent of large, consumer-centric enterprises.”* At Microsoft, we envision a world where natural language becomes the new user interface, enabling people to do more with what they say, type and input, understanding preferences and tasks and modeling experiences based on the way people think and remember.

Logo of XOXOCOToday, we are announcing we have signed an agreement to acquire XOXCO, a software product design and development studio known for its conversational AI and bot development capabilities. The company has been paving the way in conversational AI since 2013 and was responsible for the creation of Howdy, the first commercially available bot for Slack that helps schedule meetings, and Botkit, which provides the development tools used by hundreds of thousands of developers on GitHub. Over the years, we have partnered with XOXCO and have been inspired by this work.

We have shared goals to foster a community of startups and innovators, share best practices and continue to amplify our focus on conversational AI, as well as to develop tools for empowering people to create experiences that do more with speech and language.

The Microsoft Bot Framework, available as a service in Azure and on GitHub, supports over 360,000 developers today. With this acquisition, we are continuing to realize our approach of democratizing AI development, conversation and dialog, and integrating conversational experiences where people communicate.

Over the last six months, Microsoft has made several strategic acquisitions to accelerate the pace of AI development. The acquisition of Semantic Machines in May brought a revolutionary new approach to conversational AI. In July, we acquired Bonsai to help reduce the barriers to AI development by combining machine teaching, reinforcement learning and simulation. In September, we acquired Lobe, a company that has created a simple visual interface empowering anyone to develop and apply deep learning and AI models quickly, without writing code. The acquisition of GitHub in October demonstrates our belief in the power of communities to help fuel the next wave of bot development.

Our goal is to make AI accessible and valuable to every individual and organization, amplifying human ingenuity with intelligent technology. To do this, Microsoft is infusing intelligence across all its products and services to extend individuals’ and organizations’ capabilities and make them more productive, providing a powerful platform of AI services and tools that makes innovation by developers and partners faster and more accessible, and helping transform business by enabling breakthroughs to current approaches and entirely new scenarios that leverage the power of intelligent technology.

We’re excited to welcome the XOXCO team and look forward to working with the community to accelerate innovation and help customers capitalize on the many benefits AI can offer.

*Gartner, Is Conversational AI the Only UX You Will Ever Need?, 25 April 2018

Tags: , , , ,

Posted on Leave a comment

Azure teams up with Carnegie Mellon University to bring the cloud to the Living Edge Laboratory

Edge computing is one of the greatest trends that is not only transforming the cloud market, but creating new opportunities for business and society. At its core, edge computing is about harnessing compute on a device, closest to where insights need to be realized – on a connected car, a piece of machinery, or a remote oil field, for example – so these insights come without delay and with a high-degree of accuracy. This is known as the “intelligent edge” because these devices are constantly learning, often aided by AI and Machine Learning algorithms powered by the intelligent cloud. At Microsoft we see amazing new applications of the intelligent edge and intelligent cloud every day – and yet this opportunity is so expansive – the surface has barely been scratched.

To further advance inquiry and discovery at the edge, today we are announcing Microsoft is donating cloud hardware and services to Carnegie Mellon University’s Living Edge Laboratory. Carnegie Mellon University is recognized as one of the leading global research institutions. Earlier this year, the university announced a $27.5 million semiconductor research initiative to connect edge devices to the cloud. Today’s announcement builds on existing commitments to discovery in the field of edge computing.

The Living Edge Laboratory is a testbed for exploring applications that generate large data volumes and require intense processing with near-instantaneous response times. The lab is designed to open accessibility to the latest innovations in edge computing and advance discovery for edge computing applications across industries. Microsoft will donate an Azure Data Box Edge, Azure Stack partnering with Intel, and Azure credits to do advanced machine learning and AI at the edge.

This new paradigm of cloud computing requires consistency in how an application is developed, to run in the cloud and at the edge. We are building Azure to make this possible. From app platform to AI, security and management, our customers can architect, develop, and run a distributed application as a single, consistent environment.

We offer the most comprehensive portfolio to enable computing at the edge, covering the full spectrum IoT devices and sensors to bringing the full power of the cloud to the edge with Azure Stack. With this spectrum spanning hardware and software, with security and advanced analytics and AI services, Microsoft brings a robust platform for developers to research and create new applications for the edge.

The Living Edge Lab was established over the past year through the Open Edge Computing Initiative, a collective effort dedicated to driving the business opportunities and technologies surrounding edge computing. With the addition of Microsoft products to the lab, faculty and students will be able to use them to develop new applications and compare their performance with other components already in place in the lab. As part of this donation, Microsoft is also joining the Open Edge Computing Initiative.

Students of Carnegie Mellon are already making exciting discoveries and applications powered by Azure AI and ML services at the edge. One of these applications is designed to help visually impaired people to detect objects or people nearby. The video feeds of a stereoscopic camera on a user are transmitted to a nearby cloudlet, and real-time video analytics is used detect obstacles. This information is transmitted back to the user and communicated via vibro-tactile feedback.

Another, OpenRTiST, allows a user to see the world around them in real time, “through the eyes of an artist.” The video feed from the camera of a mobile device is transmitted to a local application, transformed there by a deep neural network trained offline to learn the artistic features of a famous painting, and returned to the user’s device as a video feed. The entire round trip is fast enough to preserve the illusion that the world around the user as displayed on the device is being continuously repainted by the artist.

I encourage you to learn more about Microsoft’s vision for an intelligent cloud and intelligent edge.

This is the beginning of an exciting new chapter of research at Carnegie Mellon – stemming from a collaboration on edge computing that began 10 years ago –  and we cannot wait to see what new discoveries and scenarios come to life from the Living Edge Lab.

Posted on Leave a comment

Podcast: Hearing in 3D with Dr. Ivan Tashev

Ivan Tashev podcast

Partner Software Architect, Dr. Ivan Tashev

Episode 50, November 14, 2018

After decades of research in processing audio signals, we’ve reached the point of so-called performance saturation. But recent advances in machine learning and signal processing algorithms have paved the way for a revolution in speech recognition technology and audio signal processing. Dr. Ivan Tashev, a Partner Software Architect in the Audio and Acoustics Group at Microsoft Research, is no small part of the revolution, having both published papers and shipped products at the forefront of the science of sound.

On today’s podcast, Dr. Tashev gives us an overview of the quest for better sound processing and speech enhancement, tells us about the latest innovations in 3D audio, and explains why the research behind audio processing technology is, thanks to variations in human perception, equal parts science, art and craft.

Related:


Episode Transcript

Ivan Tashev: You know, humans, they don’t care about mean square error solution or maximum likelihood solution, they just want the sound to sound better. For them. And it’s about human perception. That’s one of the very tricky parts in audio signal processing.

Host: You’re listening to the Microsoft Research Podcast, a show that brings you closer to the cutting-edge of technology research and the scientists behind it. I’m your host, Gretchen Huizinga.

Host: After decades of research in processing audio signals, we’ve reached the point of so-called performance saturation. But recent advances in machine learning and signal processing algorithms have paved the way for a revolution in speech recognition technology and audio signal processing. Dr. Ivan Tashev, a Partner Software Architect in the Audio and Acoustics Group at Microsoft Research, is no small part of the revolution, having both published papers and shipped products at the forefront of the science of sound.

On today’s podcast, Dr. Tashev gives us an overview of the quest for better sound processing and speech enhancement, tells us about the latest innovations in 3D audio, and explains why the research behind audio processing technology is, thanks to variations in human perception, equal parts science, art and craft. That and much more on this episode of the Microsoft Research Podcast.

Host: Ivan Tashev, welcome to the podcast.

Ivan Tashev: Thank you.

Host: Great to have you here. You’re a Partner Software Architect in the Audio and Acoustics groups at Microsoft Research, so, in broad strokes, tell us about your work. What gets you up in the morning, what questions are you asking, what big problems are you trying to solve?

Ivan Tashev: So, in general, in Audio and Acoustics Research Group, we do audio signal processing. That includes enhancing of a captured sound by our microphones, better sound reproduction using binaural audio, so-called spatial audio. We do a lot of work in audio analytics, recognition of audio objects, recognition of the audio background. We design a lot of interesting audio devices. Our research ranges from applied research related to Microsoft products to a blue-sky research far from what is Microsoft business today.

Host: So, what’s the ultimate goal? Perfect sound?

Ivan Tashev: Hhhh… Perfect sound is a very tricky thing, because it is about human perception. And this is very difficult to be modeled using mathematical equations. So, the classic statistical signal processing was established in 1947 with a paper published by Norbert Wiener defining what we call, today, the Wiener Filtering. The approach is simple: you have a process, you make a statistical model, you define optimality criterion, make the first derivative, make it zero, voila! You have the analytical solution of the problem. The problem is that, you either have an approximate model, and find the solution analytically, or you have precise model which you cannot solve analytically. The other thing is the optimality criterion. You know, humans, they don’t care about mean square error solution or maximum likelihood solution, they just want the sound to sound better. For them. And it’s about human perception. That’s one of the very tricky parts in audio signal processing.

Host: So, where are we heading in audio signal processing, in the era of machine learning and neural networks?

Ivan Tashev: The machine learning and neural networks are capable to find the solution from the data without us making an approximate model. And this is the beauty of this whole application of machine learning in signal processing, and the reason why we achieve significantly better results than using statistical signal processing. Even more, we train the neural network using certain cost function and we can make the cost function to be even another neural network, trained on human perception for better audio which allows us to achieve better perception of a higher quality of the speech enhancement we do using neural network. I’m not saying that we should go in every single audio processing block using machine learning and neural networks. We have processing blocks which have a nice and clean analytical solution, and this runs fast and efficient, and they will remain the same. But in many cases, we operate with approximate models with not very natural optimality criteria. And then, this is where the machine learning shines. This is where we can achieve much better results and provide a higher quality of our output signal.

Host: One interesting area of research that you are doing is noise robust speech recognition. And this is where researchers are working to improve automatic speech recognition systems. So, what’s the science behind this and how are algorithms helping to clean up the signal?

Ivan Tashev: We are witnessing a revolution in speech recognition. The classic speech recognizer was based on so-called Hidden Markov Models or HMM’s. And they served us quite well, but the revolution came when neural networks were implemented and trained to do speech recognition. My colleagues in the speech research group were the first to design a neural network-based speech recognition algorithm which instantly showed better results than the existing production HMM-based speech recognizer. The speech recognition engine has one channel input, while in audio processing, we can deal with multiple channels, so-called microphone arrays, and they give us a sense of spatiality. We can detect the direction where the sounds came from. We can enhance that sound. We can suppress sounds coming from other directions. And then provide this cleaner sound to the speech recognition engine. The microphone reprocessing technologies combined together with techniques like sound source localization and tracking and sound source separation allow us to even separate two simultaneously speaking humans in the conference room and feed two separate instances of the speech recognizer for meeting transcription.

Host: Are you serious?

Ivan Tashev: Yes, we can do that. Even more, the audio processing engine has more prior information. For example, the signal we send to the loudspeakers. And the goal of this engine is to remove the sound which is interfering for our sound. And this is also one of the oldest signal processing algorithms and every single speaker phone has it. But, in all instances, it has been implemented as a mono acoustic echo cancellation. In Microsoft, we were the first to design a stereo and surround sound echo canceller despite a paper written by the inventor of the acoustic echo cancellation himself, stating that stereo acoustic cancellation is not possible. And it’s relatively simple to understand: you have two channels between the left and the right speaker coming to one microphone, so you have one equation and two unknowns. And Microsoft released, as part of Kinect for Xbox, a surround sound echo cancellation engine. Not that we solved five unknowns from one equation, but we just found a workaround which was good enough for any practical purposes and allowed us to clean the surround sound coming from the Xbox to provide a cleaner sound to the speech recognition engine.

Host: So, did you write a paper and say, “Yes, it is possible, thank you very much!”?

Ivan Tashev: I did write a paper.

Host: Oh, you did!

Ivan Tashev: And it was rejected with the most crucial feedback from the reviewers I have ever seen in my career. It is the same to go to the French Academy of Sciences and to propose eternal engine. They have decided, since 18th century, not to discuss papers about that. When I received the rejection notice, I went downstairs in my lab, started the demo, listened to the output. Okay, it works! So, we should be fine!

(music plays)

Host: One thing that’s fascinated me about your work is the infamous anechoic chamber – or chambers, as I came to find out – at Microsoft, and one’s right here in Building 99, but there are others. And so, phrases like “the quietest place on earth” and “where sound goes to die” are kind of sensational, but these are really interesting structures and have really specific purposes which I was interested to find out about. So, tell us about these anechoic, or echo-free, chambers. How many are there here, how are they different from one another and what are they used for?

Ivan Tashev: So, the anechoic chamber is just a room insulated from the sounds outside. In our case, it’s a concrete cube which does not touch the building and sits on around half a meter of rubber to prevent vibrations from the street to come into the room. And internally, the walls, the ceiling and the floor are covered with sound absorption panels. This is pretty much it. What happens is that the sound from the source reaches the microphone, or the human ear, only using the direct path. There is no reflection from the walls and there is no other noise in the chamber. Pretty much that anechoic chamber simulates absence of a room. And it’s just an instrument for making acoustical measurements. What we do in the chamber is we measure the directivity patterns of microphones or radiation patterns of loudspeakers as they are installed in the devices we design. Initially, the anechoic chamber here, in Microsoft Building 99, the headquarters of Microsoft Research, was the only one in Microsoft. But with our engagement with product teams, it became overcrowded, and our business partners decided to build their own anechoic chambers. And there are, today, five in Microsoft Corporation. They all can perform the standard set of measurements, but all of them are a little bit different from each other. For example, the “Quietest Place in the Earth,” as recorded in the Guinness Book of Records, is the anechoic chamber in Building 88. And the largest anechoic chamber is in Studio B which allows making measurements with lower frequencies than in the rest of the chambers. In our chamber, in Building 99, it’s the only one in Microsoft which can allow human beings to stay prolonged amount of time in the chamber because we have air-conditioning connected to the chamber. It’s a different story how much effort it cost us to make the rumbling noise from the air conditioner not to enter the anechoic chamber. But this allowed us to do a lot of research on human spatial hearing in that chamber.

Host: So, drill in on that a little bit because, coming from a video production background, the air conditioner in a building is always the annoying part for the sound people. But you’ve got that figured out in the way that you situated the air conditioning unit and so on?

Ivan Tashev: To remove this rumbling sound from the air conditioner, we installed a gigantic filter which is under the floor of the entire equipment room. So, think about six by four meters floor and this is how we were able to reduce the sound from the air conditioning. Still, if you do a very precise acoustical measurement, we have the ability to switch it off.

Host: Okay. So, back to what you had said about having humans in this room for prolonged periods of time. I’ve heard that your brain starts to play tricks on you when you are in that quiet of a place for a prolonged period of time. What’s the deal there?

Ivan Tashev: OK. This is the human perception of the anechoic chamber. Humans, in general, are, I would say two and a half dimensional creatures. When we walk on the ground, we don’t have very good spatial hearing, vertically. We do much better horizontally. But also, we count on the first reflection from the ground to use it as a distance cue. When you enter the anechoic chamber, you subconsciously swallow, and this is a reaction because your brain thinks that there is a difference in the pressure between your inner ear and the atmosphere which presses the ear drums and you cannot hear anything.

Host: So that swallowing reaction is what you do when you’re in an airplane and the pressure actually changes. And you get the same perception in this room, but the pressure didn’t change.

Ivan Tashev: Exactly. But the problem in the room is that you cannot hear anything just because there is no sound in the chamber. And the other thing what happens is you cannot hear that reflection from the floor which is basically very hard-wired in our brains. We can distinguish two separate sounds when the distance between them is a couple of milliseconds. And when the sound source is far away, this difference between the direct path and the reflection from the ground is less than that. We hear this as one sound. We start to perceive those two as separate sounds when the sound source is closer than a couple of meters away… means two jumps. Then subconsciously alarm bells start to ring in our brain that, hey, there is a sound source less than two jumps away, watch out not to become the dinner! Or maybe this is the dinner!

Host: So, the progress, though, of what your brain does and what your hearing does inside the chamber for one minute, for ten minutes, what happens?

Ivan Tashev: So, there is no sound. And, the brain tries to acquire as much information as possible. And the situation when you don’t get information is called information deprival. You, first after a minute or so, start to hear a shhhhhh, which is actually the blood in the vessels of your ear. Then, after a couple of minutes, you start to hear your body sounds, your heartbeat, your breathing. And, under no other senses, eyes closed, no sound coming, literally you reach, after ten, fifteen minutes the stage of audio hallucinations. Our brains are pattern-matching machines, so sooner or later, the brain will start to recognize sounds you have heard somewhere in different places. We – people from my team – we have not reached that stage, simply because when you work there, the door is open, the tools are clanking, we have conversations, etcetera, etcetera. But maybe someday I will have to lay there and close my eyes and see, can I reach the hallucination stage?

(music plays)

Host: Well, let’s talk about the research behind Microsoft Kinect. And that’s been a huge driver of innovations in this field. Tell us how the legacy of research and hardware for Kinect led to progress in other areas of Microsoft.

Ivan Tashev: Kinect introduced us to new modalities in human-machine interfaces: voice and gesture. And it was a wildly successful product. Kinect entered the Guinness Book of Records for the fastest-selling electronic device in the history of mankind. Microsoft sold eight million devices in the first three months of the beginning of the production. Since then, most of the technologies in Kinect have been further developed. But even during the first year of Kinect, Microsoft released Kinect for Windows which allowed researchers from all over the globe to do things we even didn’t thought of. This is so-called Kinect Effect. We had more than fifty start-ups building their products using technologies from Microsoft Kinect. Today, most of them are further developed, enhanced, and are part of our products. I’ll give just two examples. The first is HoloLens. The device does not have a mouse or keyboard and the human-machine interface is built on three input modalities: gaze, gesture and voice. In HoloLens, we have a depth camera, quite similar to the one in Kinect, and we do gesture recognition using super-refined and improved algorithms, but they originate from the ones we had in Kinect. The second example is also HoloLens. HoloLens has four microphones, the same number as Kinect, and I would say that the audio enhancement pipeline for getting the voice of the person wearing the device is the granddaughter of the audio pipeline released in Kinect in 2010.

Host: Now let’s talk about one of the coolest projects you are working on. It’s the spatial audio or 3D audio. What’s your team doing to make the 3D audio experience a reality?

Ivan Tashev: In general, spatial audio or 3D audio is a technology that allows us to project audio sources in any desired position to be perceived by the human being wearing headphones. This technology is not something new. Actually, we have instances of it in mid-19th century, when two microphones and two rented telephone lines were used for stereo broadcasting of a theatrical play. Later, in the 20th century, there have been vinyl records marked to be listened with headphones because they were stereo recorded using a dummy head with two microphones in the ears. This technology did not fly because of two major deficiencies. The first is, you move your head left and right and the entire audio scene rotates with you. The second is that your brain may not exactly like the spatial cues coming from the microphones in the ear of the dummy head. And this is where we reach the topic of head-related transfer functions. Literally, if you have a sound source somewhere in the space, the sound from it reaches your left and right ear in a slightly different way. It can be modeled as two filters. And if you filter it through those two filters and play it through headphones, your brain will perceive the sound coming from that direction. If we know those pairs of filters for all directions around you, this is called head-related transfer functions. The problem is that they are highly individual. Head-related transfer functions are formed by the size and the dimensions of the head, the position of the ears, the fine structure of the pinna, the reflections from the shoulders. And we did a lot of research to find the way to quickly generate personalized head-related transfer functions. We put, in our anechoic chamber, more than four hundred subjects. We measured their HRTFs. We did a submillimeter precision scan of their head and torso, and we did measurement of certain anthropometric dimensions of those subjects. Today, we can just measure several dimensions of your head and generate your personalized head-related transfer function. We can do this even from a depth picture. Literally, you can tell how you hear from the way you look. And we polished this technology to extend that in HoloLens, you have your spatial audio personalized without even knowing it. You put the device on and you hear through your own personalized spatial hearing.

Host: How does that do that automatically?

Ivan Tashev: Silently, we measure certain anthropometrics of your head. Our engineering teams, our partners, decided that there should not be anything visible for generation of those personalized spatial hearing.

Host: So, if I put this on, say the HoloLens headset, it’s going to measure me on the fly?

Ivan Tashev: Mmm hmmm.

Host: And then the 3D audio will happen for me. Both of us could have the headset on and hear a noise in one of our ears that supposedly is coming from behind us, but really isn’t. It’s virtual.

Ivan Tashev: That’s absolutely correct. With the two loudspeakers in HoloLens or in your headphones, we can make you perceive the sound coming from above, from below, from behind. And this is actually the main difference between surround sound and 3D audio for headphones. Surround sound has five or seven loudspeakers, but they are all in one plane. So, surround audio world is actually flat. While with this spatial audio engine, we can actually render audio above and below which opens pretty much a new frontier in expressiveness of the audio, what we can do.

Host: Listen, as you talk, I have a vision of a bat in my head sending out signals and getting signals and echolocations and…

Ivan Tashev: We did that.

Host: What?

Ivan Tashev: We did that!

Host: Okay, tell.

Ivan Tashev: So, one of our projects – this is one of those more blue-sky research projects – is exactly about that. What we wanted to explore is using audio as echolocation in the same way the bats see in complete darkness. And we built a spherical loudspeaker array of eight transducers which sent ultrasound pulses towards given direction, and near it, an eight-element microphone array which, through the technology called beam forming, listens towards the same direction. With this, we utilized the energy of the loudspeakers well, and reduced the amount of sounds coming from other directions and this allows us to measure the energy reflected by the object in that direction. When you do the scanning of the space, you can create an image which is exactly the same as created from a depth camera using infrared light but with a fraction of the energy. The ultimate goal, eventually, will be to get the same gesture recognition with one tenth or one hundredth of the power necessary. This is important for all portable battery-operated devices.

Host: Yeah. Speaking of that, accessibility is a huge area of interest for Microsoft right now, especially here in Microsoft Research with the AI for Accessibility initiative. And it’s really revolutionizing access to technology for people with disabilities. Tell us how the research you’re doing is finding its way into the projects and products in the arena of accessibility.

Ivan Tashev: You know, accessibility finds a resonance among Microsoft employees. The first application of our spatial audio technology was actually not HoloLens. It was a project which was a kind of a grass roots project when Microsoft employees worked with a charity organization called Guide Dogs in United Kingdom. And from the name you can basically guess that they train guiding dogs for people with blindness. The idea was to use the spatial audio to help the visually impaired. Multiple teams in Microsoft Research, actually, have been involved to overcome a lot of problems, including my team, and this whole story ended up with releasing a product called Soundscape, which is a phone application which allows people with blindness to navigate easier where the spatial audio acts like a finger-pointer. When the system says, “And on the left is the department store,” actually that voice-prompt came from the direction where the department store is, and this is additional spatial cue which helps the orientation of the visually impaired people. Another interesting project we have been involved, also is a grass roots project. It was driven by a girl which was hearing-impaired. She initiated a project during one of the yearly hackathons. And the project was triggered by the fact that she was told by her neighbor that your CO2 alarm is beeping already a week. You have to replace the battery. So, we created a phone application which was able to recognize numerous sounds like CO2 alarm, fire alarm, door knock, phone ring, baby crying, etcetera, etcetera, and to signal the hearing-impaired person using vibration, or the display. And this is to help to navigate and to live a better life in our environment.

(music plays)

Host: You have an interesting personal story. Tell us a bit about your background. Where did you grow up, what got you interested in the work you are doing and how did you end up at Microsoft Research?

Ivan Tashev: I’m born in a small country in Southeast Europe called Bulgaria. I took my diploma in electronic engineering, and PhD in computer science from the Technical University of Sofia, and immediately after my graduation, started to work as a researcher there. In 1998, I was Assistant Professor in the Department of Electronic Engineering when Microsoft hired me, and I moved to Washington State. Spent to two full shipping cycles in Microsoft engineering teams before, in 2001, to move in Microsoft Research. And what I have learned during those two shipping cycles actually helped me a lot to talk better with the engineers during the technology transfers I have done with Microsoft engineering teams.

Host: Yeah, and there’s quite a bit of tech transfer that’s coming out of your group. What are some examples of the things that have been “blue sky research” at the beginning that are now finding their way into millions of users’ desks and homes?

Ivan Tashev: I have been lucky enough to be a part of very strong research groups and to learn from masters like Anoop Gupta or Rico Malvar. My first project in Microsoft Research was called Distributed Meetings and we used that device to record meetings, to store them and to process them. Later, this device became a roundtable device which is part of many conference rooms worldwide. Then, I decided to generalize the microphone array support I designed for round table device and this became the microphone array support in Windows Vista. Next challenge was to make this speech enhancement pipeline to work even in more harsh conditions like the noisy car. And, I designed the algorithms and transferred them to the first speech-driven entertainment system in a mass-production car. And then, the story continues with Kinect, with HoloLens, many other products, and this is another difference between industrial research and academia. The satisfaction from your work is measurable. You know to how many homes your technology has been released, to how many people you changed the way they live, entertain or work.

Host: As we close, Ivan, perhaps you can give some parting advice to those of our listeners that might be interested in the science of sound, so to speak. What are the exciting challenges out there in audio and acoustics research, and what guidance would you offer would-be researchers in this area?

Ivan Tashev: So, audio processing is a very interesting area of research because it is a mixture between art, craft and science. It is science because we work with mathematical models and we have repetitive results. But it is an art because it’s about human perception. Humans have their own preferences and tastes, and this makes it very difficult to model with mathematical models. And it’s also a craft. There are always some small tricks and secret sauce which are not mathematical models but make the algorithms from one lab work much better than the algorithms from another lab. Into the mixture, we have to add the powerful innovation of machine learning technologies, neural networks and artificial intelligence which allow us to solve problems we thought were unsolvable and to produce algorithms which work much better than the classic ones. So, the advice is, learn signal processing and machine learning. This combination is very powerful!

Host: Ivan Tashev, thank you for joining us today.

Ivan Tashev: Thank you.

To learn more about Dr. Ivan Tashev and how Microsoft Research is working to make sound sound better, visit Microsoft.com/research.

Posted on Leave a comment

‘Fallout 76’ now available on Xbox one

The online prequel to the epic Fallout series, Fallout 76 throws you into a multiplayer-filled wasteland where you all must work together (or not) to survive the constant threat of nuclear annihilation, all the while experiencing the largest, most dynamic open world ever created in the legendary Fallout universe. Brought to you by the award-winning creators of Skyrim and Fallout 4, Fallout 76 is now available on Xbox One.

Fallout 76 Screenshot

Fallout 76 Screenshot

A Player-Filled Wasteland

Forge your own path in a wild wasteland with tons of new and unique locations to discover. With a refined S.P.E.C.I.A.L. system that lets you craft your own personality in this new world, you will have the option to journey alone or with friends.

Fallout 76 Screenshot

Fallout 76 Screenshot

Almost Heaven

With all-new graphics, lighting, and landscape tech, six distinct West Virginia regions will be brought to life in Fallout 76. From the thick, green forests of Appalachia to the toxic expanse of the Cranberry Bog, each unique region will offer its own risks and rewards — post-nuclear America has never looked so good.

Fallout 76 Screenshot

Fallout 76 Screenshot

Go C.A.M.P.ing

Introducing the all-new Construction and Assembly Mobile Platform (C.A.M.P.). With the C.A.M.P. you can build and craft anywhere in the world of Fallout 76. Use it to create much-needed shelter, safety, and supplies to survive the West Virginia wasteland. You can also use the C.A.M.P. to trade with other survivors (players). Just keep a close eye on neighbors who might be a little too friendly…

Fallout 76 Screenshot

Fallout 76 Screenshot

Get Contaminated

Part of your adventure may lead you to uncover the ultimate weapon in Fallout 76: Nuclear Missiles. Now you can be part of the destruction of humanity when the bombs fall, where the destruction will leave behind a high-level zone with rare and valuable resources. It’ll be up to you to unleash or protect the power of the atom.

Fallout 76 Screenshot

Fallout 76 Screenshot

Fallout 76 is available now on the Microsoft Store for the Xbox One family of devices and is an Xbox One X Enhanced title. You can also pick up the Fallout 76: Tricentennial Edition that includes bonus in-game digital items like Tricentennial Power Armor, a Vault Boy Mascot Head, Patriotic Uncle Sam Outfit, and more. Stay tuned to Xbox Wire for all of the latest news on your favorite Xbox One games as well as the latest news and updates on Fallout 76.

Posted on Leave a comment

Microsoft driving intelligent, connected smart cities

Barcelona may be one of the world’s major global centers for tourism, the economy, trade and culture but, starting today, it’s the apex for smart cities. Our Microsoft CityNext team and partners are excited to join government and civic leaders as well as experts at the Smart City Expo World Congress (SCEWC2018) to address key issues facing cities, including:

  • Enhancing urban mobility
  • Building smarter infrastructure
  • Improving citizen and social care
  • Strengthening public safety.

Innovations for smart cities

While the Internet of Things (IoT) has been part of smart cities strategy from the early days, we are now reaching a tipping point where these solutions can be created at scale and securely, using sophisticated models and artificial intelligence capabilities to improve our work and personal lives across these scenarios. Microsoft is bringing its leading capabilities from the intelligent cloud to the intelligent edge into these solutions.

This approach for creating intelligent, connected smart cities results in digital transformations and services that address priorities, better serve citizens, and enable more sustainable, prosperous and inclusive communities. For example, the convergence of the cloud, AI and IoT is behind our new Azure Digital Twins service, which allows cities and other customers to model the relationships and interactions among people, places and devices before connecting devices to that model. Learn more about Digital Twins here.

During the recent IoT Solutions World Congress, we announced that IoT has the potential to create more efficient and vibrant cities by providing new insights and approaches to transportation and traffic, energy reduction, construction, utilities, parking and much more. Learn how our IoT solutions are enabling transformations here.

In another indication that the Azure cloud continues gaining momentum, Moovit, the world’s largest urban mobility data and analytics company and No. 1 transit app, announced last week that it will integrate its sophisticated API data into Azure Maps. This partnership provides a comprehensive mapping solution that brings together location awareness with public transit data, enabling developers to build richer apps requiring public transportation routing services—such as city planning around public transit, multi-modal routing and HOV or toll-road avoidance route optimization—helping cities harness intelligence to minimize urban congestion and reduce their carbon footprint.

Cities transforming with the cloud, AI and IoT

During SCEWC2018, Microsoft will join city leaders, solution experts and partners to showcase modern solutions built on our cloud, AI and IoT. These powerful technologies are enabling “intelligent cities” across the U.S. and around the world, for example:

  • Antwerp, Belgium, chose Microsoft Azure-based Be-Mobile technology to build its SlimNaarAntwerp platform, which is providing citizens with a mobility-as-a-service solution to identify an optimum trip, combining different means of transportation—car, train, shared bikes and on foot. The plan is to avoid 20,000 car movements to Antwerp’s city center during rush hours while improving satisfaction.
  • Denver is aggressively embracing and testing new and better ways to deploy technology and using data to improve services for residents, businesses and visitors. Building on Azure and other Microsoft technologies, the Mile-High City is in the middle of a four-year, $12 million smart city program to reduce traffic congestion, improve pedestrian safety and enhance quality of life, particularly in underserved communities.
  • Espoo, Finland, is a hub of technology and business innovation, and the city wants to deliver public services that are just as innovative. To better understand the service needs of its 282,000 citizens, Espoo together with software and services company Tieto built an experimental City-as-a-Service platform that combines Tieto’s AI and advanced analytics know-how with cloud services from Microsoft Azure. A challenge is how to offer the right services in the right time for the right customers in predictive ways. To do that, Espoo needed a platform that could process more than 500 million rows of data—from social and healthcare, daycare services and other agency records—at any one time. The successful experiment demonstrated that the city can use the platform to predict which citizens may need social welfare services, provide services proactively—and deliver far better outcomes.
  • Houston is partnering with Microsoft to drive the city forward with technology innovations that address key priorities, including disaster recovery and response, building and school safety, and more efficient, capable transportation, as shown in this video. For example, using our cloud-powered IoT hub, AI and Cognitive Services, Houston is deploying a connected buses solution to provide internet access to riders and increase safety by: installing a driver panic button to get help with onboard medical emergencies or altercations, a fleet management system to confirm buses are being driven safely and an alert to identify near-term maintenance requirements that help prevent breakdowns.

Partners joining us at SCEWC2018

These 12 CityNext partners are joining us in Barcelona to showcase solutions built on our platform to support smart city initiatives and modernization projects around the world:

Connect with us—at SCEWC2018 or your city

Microsoft CityNext looks forward to connecting with city leaders attending SCEWC2018 this week. We invite you to join us at the Intelligent Cities Forum at Smart City Expo, hosted in the East Agora from 4 to 7 p.m., Tuesday, Nov. 13. During this fast-paced agenda, city leaders, solution experts and Microsoft executives will showcase​ industry-leading, modern technology to enable intelligent cities. We’ll highlight areas such as smart infrastructure, urban mobility, social care and public safety in the context of two critical scenarios. This forum will focus on the “how” of building an intelligent city and provide you with real examples to ​implement today. Tapas and refreshments provided.

We also welcome the opportunity to help your city find its path to digital transformation with our cloud, AI and IoT solutions. To get started, please utilize these additional resources:

Learn how AI is transforming industries around the world by reading the Intelligent Economies: AI’s Transformation of Industries and Societies Report by the Economist Intelligence Unit.