Home Blog Page 341

China Overtakes The US In AI Journal Citation – Stanford AI Index Report

Stanford AI Index Report

Stanford publishes its AI Index Report that focuses on the developments of the complex artificial intelligence landscape since 2017. The latest report — 2021 — shed some light on the impact of COVID-19 in AI research, countries leading the race in research, and more. In a total of 7 chapters, the Stanford AI Index Report also covers aspects like AI education, research and development, diversity in AI, and AI policy.

One of the most surprising revelations, for many, is that China overtakes the US in terms of journal citation, pinpointing the advancement in their research. This comes after China surpassed the US in the terms of the number of artificial intelligence research publications in 2017 after briefly overtaking in 2004. However, the US has significantly more cited AI conference papers than China.

While experts and AI enthusiasts may be at crossroads when it comes to the development of research breakthroughs from China, everyone can be blissful about the expanding research in artificial intelligence. According to Stanford AI Index Report, in 2020, the artificial intelligence journal publications grew by 34.5 percent from 2019. AI-related research publications also grew sixfold on arXiv since 2015 and reached 34,736 in 2020.

Also Read: MIT Task Force: No Self-Driving Cars For At Least 10 Years

The COVID-19 pandemic also provided the opportunity for machine learning enthusiasts and practitioners to attend more virtual conferences; attendance across nine AI-related conferences almost doubled in 2020. The interest among aspirants to build the real-world application using artificial intelligence is also evident, as, in 2019, 65 percent of PhDs from North America moved into industry, up from 44.4 percent in 2010.

While there was a lot to cheer about this report, some of the challenges the AI industry witnessing are startling. One of the major challenges that are impeding AI development is the need for diversity. The Stanford AI Index Report pinpointed that, in 2019, 45 percent of new U.S. resident AI Ph.D. graduates were white, 2.4 percent were African American, and 3.2 percent Hispanic.

You can read about the entire report here, which has a lot of information on the issues with the lack of proper benchmarking for artificial intelligence to measure the real-world impact and more.

Advertisement

NVIDIA Rolls Out Transfer Learning Toolkit 3.0: Build Production-Ready ML Models Without Coding

NVIDIA Transfer Learning Toolkit

The largest GPU provider releases NVIDIA Transfer Learning Toolkit 3.0 to help professionals build production-quality computer vision and conversational AI models without coding. As the name suggests, the toolkit leverages the transfer learning technique — a method where a deep learning model transfers its learning to another model to further improve the dexterity of the newer models.

To develop a deep learning model, data scientists need superior computation, large-scale data collection, labeling, statistics, maths, model development expertise, and more. This impedes practitioners from quickly developing machine learning models and bring them to the market. Consequently, NVIDIA rolls out the Transfer Learning Toolkit 3.0 to eliminate the need for a wide range of expertise to build exceptional models.

embedded-diagram-tlt3.0-launch-software-stack-diagram-1571862-01.png
NVIDIA Transfer Learning Toolkit

Data scientists can now develop models by just fine-tuning the pre-trained models from the NVIDIA Transfer, thereby building models even without having a knowledge of artificial intelligence frameworks. According to the company, the NVIDIA Transfer Learning Toolkit (TLT) can expedite the engineering efforts by 10x. In other words, deep learning models that usually take 80 weeks while building from the ground up, with NVIDIA TLT, development can be carried out in 8 weeks.

These pre-build models are available for free and can be accessed from NGC to develop common computer vision and conversational AI models like people detection, text recognition, image classification, license plate detection and recognition, vehicle detection and classification, facial landmark, heart rate estimation, and more.

All you have to do to get started is pull the NVIDIA Transfer Learning Toolkit container from NGC, which comes pre-packaged with Jupyter Notebooks, to build state-of-the-art models. Although the Toolkit is available for commercial use, you should check for the specific terms of license of models

Read more here and get started with tutorials here.

Advertisement

NVIDIA Deep Learning Institute Introduces Free Accelerated Data Science Teaching Kit

NVIDIA accelerated data science teaching kit

NVIDIA Deep Learning Institute that offers free hands-on training in artificial intelligence, has released a new Data Science Teaching Kit. The idea behind NVIDIA accelerated data science teaching kit is to help learners gain access to high performance computing, exceptional libraries, and other machine learning techniques. 

The accelerated data science teaching kit is packed with resources for fundamentals and advanced topics in data collection and processing accelerated data science with RAPIDS, GPU-accelerated machine learning, data visualization, data ethics and bias in datasets, data visualization, and more.

The NVIDIA accelerated data science teaching kit is devised in collaboration with two experts — Polo Cahu and Xishung Dong — of the Georgia Institute of Technology and Prairie View A&M University. The teaching kit has lecture slides, notes, quizzes, and other hands-on labs. And according to NVIDIA, in the future, videos will be released for the modules to allow an omnichannel learning experience.

Also ReadInfosys Cobalt Announces Applied AI Cloud Built On NVIDIA DGX A100s

Over time, NVIDIA Deep Learning Institute has released teaching kits that come with free GPU with AWS credits for educators and their students, self-paced courses and certificate opportunities, live instructor-led workshops, and more. With the release of the accelerated data science teaching kit, Deep Learning Institute has not released a total of 4 teaching kits for educators and their learners.

Currently, teaching kit is available in categories like Accelerated Computing, Data Science, Deep Learning, and Robotics. By gaining access to these teaching kits for free, learners can obtain knowledge of applying machine learning techniques and build end-to-end projects with ease. However, the toolkit is only available for qualified educations. You will have to apply to get access to the free toolkit. A wide range of people, including students, professors, and researchers, can join in by applying for the permission of Deep Learning Institute teaching kit.

Click here to apply for the NVIDIA teaching kit and learn the latest technologies for free.

Advertisement

Google Introduces Model Search, An Open-Source Platform To Find The Best ML Model

Google Model Search

Data scientists struggle to find the best model for their projects as there can be many factors that can influence the performance of machine learning models. To mitigate such challenges, Google introduces Model Search — a framework to implement AutoML algorithms for model architecture search at scale. Just like any other machine learning practitioner, if you come across questions like ‘which is the appropriate neural network should be implemented?’ ‘LSTMs or Transformer?’ ‘Ensembling or distillation for performance?’ and more, the library is for you to find the right answers.  

Google Model Search library will allow data scientists to run AutoML algorithms on their data to find the best model with the right layers for the project. What makes Google Model Search superior is that it considers which domain the project is catering to for finding the best model architecture.

Developers can get started with the model search with only a few lines of code, and the library can run hundreds of machine learning models. Post checking with several models data scientists can check the results of individual models’ performance in the root directory. There are default specifications that are used while evaluating numerous models on data but developers can create their own specifications as well. Besides, Google Model Search also enables developers to test their own models — called blocks.

Google has also ensured that the search can run parallelly to reduce the turnaround time with the help of multiple machines. However, the current version of the framework only supports classification problems only. In the future, it will also empower developers to use regression problems.

As per the researchers, Google Model Search has shown exceptional results that were demonstrated in the recent paper in the speech domain. “Over fewer than 200 iterations, the resulting model slightly improved upon internal state-of-the-art production models designed by experts in accuracy using ~130K fewer trainable parameters (184K compared to 315K parameters),” mentions the researchers in a blog post.

Find the Google Model Search library on GitHub.

Advertisement

AWS To Host A Free AI/ML Edition Conference Of AWS Innovate

aws innovate ai ml

AWS will host a free AI/ML edition of AWS Innovate for both machine learning beginners and practitioners. AWS Innovate is packed with 50+ sessions, model deployments best practices, business use cases, hands-on guides, technical demos, and more. The event will be hosted on 24 February 2021 to assist machine learning enthusiasts to learn from the experts of AWS.

Learners will also get a certificate of attendance if they completely watch 5 sessions or more of AWS Innovate AI/ML edition.

The sessions are designed into 4 levels of attendees’ expertise on various AWS and AI/ML topics. While the basic levels include imparting knowledge of getting started with AWS for AI and ML, advanced levels have sessions like optimization of recommendations, choosing the right ML algorithms for different use cases, and more.

AWS Innovate AI and ML edition also caters to startups as it has separate sessions like building AI-powered applications without any machine learning expertise, scaling your startup, among others.

Other interesting events at AWS Innovate AI/ML edition are AWS DeepRacer — a beginner’s way of getting started with reinforcement learning by training an agent for autonomous driving. Users can choose to participate in the AWS DeepRacer event and compete with learners from across the world.

Register for the AWS Innovate AI/ML virtual event here.

Advertisement

Trello Adds Visualisation To Bring Insights Into Projects

Trello Update
Image credit: Trello

As remote working is becoming the new normal, Trello releases an update to its project management tool. This major update brings several features like a dashboard, a timeline, a table, and more. 

The idea behind the new release is to offer context to the content to help teams work effortlessly while being at different places. In addition to the feature release, there is a once-in-a-decade makeover of Trello’s logo and brand.

Of the many features, visualization invoked interest among users as it offers insight into projects’ progress, assigned cards, deadlines, and others. With this update, Trello is focused on bringing new visibility into projects with ease, assisting in generating reports that can be assimilated quickly with visuals.

Built on top of AWS, Trello has over 50 million users and has plans to double down on acquiring new users with added functionalities that streamline the workflow of organizations. For this, the new Trello has added tables, which are capable of replacing spreadsheets for project management, according to the company.

To transform teams’ capabilities, Trello has also added new functionalities to views, cards, and more. Besides, calendar and due date are a new addition, where you can manage tasks and create commands to assign actions of cards based on events. For instance, the moment a card is due, move the card to the top of the list “To Do” and join the card.

Note: Several features are only available for business class and enterprise customers.

Read more here.

Advertisement

Salesforce Uses AWS Textract For Intelligent Document Automation

Salesforce AWS Textract
Credit: Salesforce

The healthcare domain has received all-time higher attention because of the current pandemic. Medical organizations have felt the heat of managing a heavier workload, mostly because of medical paperwork. Various companies have come up with solutions to reduce the person-hours in handling patient data. One of the major names, Salesforce, has been capturing the ground by offering Health Cloud as a Patient Management Service. Salesforce is using AWS Textract API to provide the Intelligent Document Automation service in its Health Cloud. Salesforce has integrated IDA into the Health Cloud to automate the manual entry of medical forms and digitize paperwork silos of the past.

The AWS Textract service is based on Optical Character Recognition (OCR) that can extract text, forms, and tables from structured documents. As of now, Salesforce service works with PDF, JPG, and PNG image files. For printed documents, English, Spanish, German, Italian, Portuguese, and French are supported.

The API detects a document’s layout and the key elements on the page, like tables and forms. It also understands the relationships between data of the embedded forms or tables and extracts everything without altering its context. Textract service was upgraded recently to enable hand-writing recognition. Currently, English is the only supported language for handwritten documents.

Read More: A Deep Dive Into IBM Quantum Roadmap

From a privacy point of view, Salesforce seems to have chosen the right API because the Textract API is compliant with Service Organization Control (SOC), International Organization for Standardization (ISO) standards, and regulations like PCI, HIPAA, and GDPR. The extracted values like patient information, such as diagnoses or prescriptions, extracted from documents used in the patient intake process are encrypted for adding another layer of security.

With an increasing number of innovations like this, Salesforce consistently emerged as a leader as accredited in The Forrester Wave™: Healthcare CRM Providers, Q1 2020, KLAS, 2018, 2019, and 2020.

Advertisement

Extracting Vocals And Instrumentals From Music The Deep Learning Way

Deep learning Vocals Instrumentals

Whenever people get exposed to good music, the tune gets stuck in their heads for hours. And at some point, they google up the lyrics, vocals, and instrumental. And the search results always point to verses uploaded by any kind-hearted people, vocals performed by individuals, and instrumentals, again performed by any person or group of people. In each case, human intervention is prevalent for optimal results. What if there is another way of getting these things?

LALAL.AI flaunts the deep-learning way; it uses state-of-the-art deep-learning models to get you the vocals and instrumentals from any music file without quality loss. It had already beaten popular vice isolation services like Spleeter by Deezer and PhonicMind regarding accessibility and quality. Neither the users have to install any program nor dive into the terminals’ darkness to get their desired materials. They simply have to drag-and-drop the music file onto their website and get their needed vocals and instrumentals.

Also Read: Facebook Releases Code Of Its State-Of-The-Art Voice Separation Model

Moreover, the LALAL.AI service always outputs files in the same format as the uploaded file, contrary to other services that only split out 44.1kHz/16bit WAV files. Users, henceforth avoid third-party services for conversion to the original format and upsampling the bitrates that further introduces noise. In a blog post, the company had put up a quantitative comparison against its competitions.

The AI company claims that it had trained extraction models with a humongous 20TB of training data. Their music dataset consists of studio-quality multi-track recordings, the same material sound engineers use. The deep-learning models behind the service have a hopping 45 million parameters. 

The LALAL.AI service can currently deconstruct remixed songs into original songs, along with their vocal and instrumental tracks. The deep learning algorithms isolate each stem precisely and hence, achieve speeds-up in track splitting. The free service offers three instances for now and for heavy users; other options are available too. They also provide an API for scalable solutions.

Advertisement

Microsoft Speller100: A Spell-Checker For Over 100 Languages

Microsoft Speller100 spell-checker

People do not care enough to use their queries’ correct spelling while searching for anything online. This recklessness makes the search engine match the incorrect set of documents and trigger wrong search results. It is of utmost importance that correctly spelled queries are submitted. Most people do not spell check because they assume that the search engines will figure out what they want to find. Thankfully, the search engines do get that right; you will find something like – – “Did you mean ‘__’?” just below their search bars. These corrections at their core are based on English. However, on a global scale, the multilingualism of the population creates new technological challenges. The linguistic diversity of the queries have quickly gone beyond mere 100 languages.

Microsoft’s search engine, Bing, which had been serving corrections in more than 24 languages, obviously had more room for improvements. Enters Speller100, the large-scale multilingual spelling correction models for more than 100 languages. It is an improvement over the traditional statistical models based on the Noisy-channel coding theorem, and user feedback on auto-correction works well for resource-heavy languages.

The researchers noted, “For a language with very little web presence and user feedback, it’s challenging to gather an adequate amount of training data. To create spelling correction solutions for these latter types of languages, models cannot rely solely on training data to learn the spelling of a language.”

Also Read: Dealing With Racially-Biased Hate-Speech Detection Models

Fundamentally, spelling correction is different from predicting the next words or sentences. So, the Speller100 needs to model both the language and the spelling errors. The spelling errors, inherently, are character level mutations. These errors have two different types – Non-word error, words out of the language vocabulary, and Word errors where the word is valid but does not fit into the context.

The error correction process was formulated as a denoising problem that converts corrupted texts to their original form. They considered the sequence-to-sequence nature of spelling and the errors as noises. All they needed was a denoising sequence-to-sequence deep learning model. Thankfully, Facebook AI already had the groundwork done with their BART paper. The Microsoft researchers leveraged the BART model that uses word-level denoising s2s autoencoder pretraining. But instead of word-level, character-level corruptions were added to terms, and an error-correcting model was trained, which shall get us back to the original word. The researchers swiftly avoided the collection of misspelled queries in 100+ languages.

The researchers had to take care of light-resource languages, where training data was not available. A zero-shot training paradigm was used, which is effective in data-scarce situations like this and does not require any additional language-specific labeled training data. They exploited the linguistic similarities of the light-resource languages with any major language family to pre-training the zero-shot models. They used resources from any related resource-heavy language. A small example has been provided below.

Microsoft claims that during the online A/B testing of the spell-checker on Bing, no-result pages reduced by 30%, user-based manual intervention for query reform went below by 5%, spellings suggestions improved by 67%, and the click-through rate of the first page went up by a staggering 70%. The company seems to ramp-up the integration of Speller100 into its other services.

Advertisement

A Deep Dive Into IBM Quantum Roadmap

IBM Quantum Roadmap.jpg
Credit: IBM Quantum

“It took us 60 years from the first logic gates to modern cloud services. But IBM has set itself on a mission to fast forward the same journey for Quantum Computation (QC) to 3 years,” Jay Gambetta, IBM Fellow and Vice President, IBM Quantum.

Quantum Computing has opened up new doors to solve existing impenetrable problems and turn them into opportunities. But realizing the promised quantum power is still “The Road not taken.” To get the computations right, one needs exactly three things — the hardware stack, the software stack, and the developer environment. IBM has been a front-runner in the quantum race by settings above three things as early as 2010. And for each component, they had released a roadmap that is a brave move, considering that corporations usually post achievements rather than work plans. The company is bold enough to take chances and announced its ultimate goal openly — “to design a full-stack quantum computer deployed via the cloud that anyone around the world can program.”

IBM Quantum Hardware Roadmap

IBM wants to build scalable, larger, and better quantum computers, and the first objective, for now, is to create a 1000+ qubit system named Condor. But the researchers have to find out the solutions to preserve the states in the qubits for a more extended period while reducing noise-induced errors to make quantum computation viable. They have been busy optimizing the compilers, continuously refining  2-qubit gates, and many more to release the next candidate, ‘Eagle,’ surpassing 100+ qubit count and concurrently processing classical computations.

Also Read: IBM And Daimler Simulates Materials With Fewer Qubits

IBM Quantum Software Roadmap

The Circuit API is still in use for sending quantum instructions to quantum computers, which can handle smaller qubit systems. But to scale things up, more powerful circuits are needed that bring iterative phase estimation closer to the qubit systems than on the users’ system and cloud. 

IBM will soon release the new Qiskit runtime to boost the performance of the hybrid cloud technology 100 times. The Dynamic Circuits are also a force-multiplier that allows branching within the circuit based on measurements. These dynamic circuits can manipulate the future states based on the intermittent measurement and produce mid-circuit resets. This ability provides support for larger algorithmic complexities and circuit variety. As the software matures, some circuits will be used more frequently than the rest, so the goal is to offer a library of pre-built and optimized circuits for end-users. 

IBM Quantum Developer Environment Roadmap

IBM has shown that developers need not learn new tools or languages but instead use their existing code to interact with a quantum computer. All they need is a few lines of code to call a quantum API service on the cloud. The company is hopeful that the developers will be able to lay the foundation for the software stack that runs on the cloud accessible by anyone worldwide. In the long run, the developers and researchers from other fields shall seamlessly integrate quantum computing into their workflow.

Apart from the corporate cut-throating for profits from quantum services, the good thing that has happened is IBM’s support for the open-source developers and researchers building the quantum applications. The company is also busy outreaching to future generations by conducting wholesome courses, giving free credits to the quantum cloud, and providing ready-to-start developing materials like its Qiskit Handbook.

Advertisement