Thursday, November 28, 2024
ad
Home Blog Page 332

Salesforce Uses AWS Textract For Intelligent Document Automation

Salesforce AWS Textract
Credit: Salesforce

The healthcare domain has received all-time higher attention because of the current pandemic. Medical organizations have felt the heat of managing a heavier workload, mostly because of medical paperwork. Various companies have come up with solutions to reduce the person-hours in handling patient data. One of the major names, Salesforce, has been capturing the ground by offering Health Cloud as a Patient Management Service. Salesforce is using AWS Textract API to provide the Intelligent Document Automation service in its Health Cloud. Salesforce has integrated IDA into the Health Cloud to automate the manual entry of medical forms and digitize paperwork silos of the past.

The AWS Textract service is based on Optical Character Recognition (OCR) that can extract text, forms, and tables from structured documents. As of now, Salesforce service works with PDF, JPG, and PNG image files. For printed documents, English, Spanish, German, Italian, Portuguese, and French are supported.

The API detects a document’s layout and the key elements on the page, like tables and forms. It also understands the relationships between data of the embedded forms or tables and extracts everything without altering its context. Textract service was upgraded recently to enable hand-writing recognition. Currently, English is the only supported language for handwritten documents.

Read More: A Deep Dive Into IBM Quantum Roadmap

From a privacy point of view, Salesforce seems to have chosen the right API because the Textract API is compliant with Service Organization Control (SOC), International Organization for Standardization (ISO) standards, and regulations like PCI, HIPAA, and GDPR. The extracted values like patient information, such as diagnoses or prescriptions, extracted from documents used in the patient intake process are encrypted for adding another layer of security.

With an increasing number of innovations like this, Salesforce consistently emerged as a leader as accredited in The Forrester Wave™: Healthcare CRM Providers, Q1 2020, KLAS, 2018, 2019, and 2020.

Advertisement

Extracting Vocals And Instrumentals From Music The Deep Learning Way

Deep learning Vocals Instrumentals

Whenever people get exposed to good music, the tune gets stuck in their heads for hours. And at some point, they google up the lyrics, vocals, and instrumental. And the search results always point to verses uploaded by any kind-hearted people, vocals performed by individuals, and instrumentals, again performed by any person or group of people. In each case, human intervention is prevalent for optimal results. What if there is another way of getting these things?

LALAL.AI flaunts the deep-learning way; it uses state-of-the-art deep-learning models to get you the vocals and instrumentals from any music file without quality loss. It had already beaten popular vice isolation services like Spleeter by Deezer and PhonicMind regarding accessibility and quality. Neither the users have to install any program nor dive into the terminals’ darkness to get their desired materials. They simply have to drag-and-drop the music file onto their website and get their needed vocals and instrumentals.

Also Read: Facebook Releases Code Of Its State-Of-The-Art Voice Separation Model

Moreover, the LALAL.AI service always outputs files in the same format as the uploaded file, contrary to other services that only split out 44.1kHz/16bit WAV files. Users, henceforth avoid third-party services for conversion to the original format and upsampling the bitrates that further introduces noise. In a blog post, the company had put up a quantitative comparison against its competitions.

The AI company claims that it had trained extraction models with a humongous 20TB of training data. Their music dataset consists of studio-quality multi-track recordings, the same material sound engineers use. The deep-learning models behind the service have a hopping 45 million parameters. 

The LALAL.AI service can currently deconstruct remixed songs into original songs, along with their vocal and instrumental tracks. The deep learning algorithms isolate each stem precisely and hence, achieve speeds-up in track splitting. The free service offers three instances for now and for heavy users; other options are available too. They also provide an API for scalable solutions.

Advertisement

Microsoft Speller100: A Spell-Checker For Over 100 Languages

Microsoft Speller100 spell-checker

People do not care enough to use their queries’ correct spelling while searching for anything online. This recklessness makes the search engine match the incorrect set of documents and trigger wrong search results. It is of utmost importance that correctly spelled queries are submitted. Most people do not spell check because they assume that the search engines will figure out what they want to find. Thankfully, the search engines do get that right; you will find something like – – “Did you mean ‘__’?” just below their search bars. These corrections at their core are based on English. However, on a global scale, the multilingualism of the population creates new technological challenges. The linguistic diversity of the queries have quickly gone beyond mere 100 languages.

Microsoft’s search engine, Bing, which had been serving corrections in more than 24 languages, obviously had more room for improvements. Enters Speller100, the large-scale multilingual spelling correction models for more than 100 languages. It is an improvement over the traditional statistical models based on the Noisy-channel coding theorem, and user feedback on auto-correction works well for resource-heavy languages.

The researchers noted, “For a language with very little web presence and user feedback, it’s challenging to gather an adequate amount of training data. To create spelling correction solutions for these latter types of languages, models cannot rely solely on training data to learn the spelling of a language.”

Also Read: Dealing With Racially-Biased Hate-Speech Detection Models

Fundamentally, spelling correction is different from predicting the next words or sentences. So, the Speller100 needs to model both the language and the spelling errors. The spelling errors, inherently, are character level mutations. These errors have two different types – Non-word error, words out of the language vocabulary, and Word errors where the word is valid but does not fit into the context.

The error correction process was formulated as a denoising problem that converts corrupted texts to their original form. They considered the sequence-to-sequence nature of spelling and the errors as noises. All they needed was a denoising sequence-to-sequence deep learning model. Thankfully, Facebook AI already had the groundwork done with their BART paper. The Microsoft researchers leveraged the BART model that uses word-level denoising s2s autoencoder pretraining. But instead of word-level, character-level corruptions were added to terms, and an error-correcting model was trained, which shall get us back to the original word. The researchers swiftly avoided the collection of misspelled queries in 100+ languages.

The researchers had to take care of light-resource languages, where training data was not available. A zero-shot training paradigm was used, which is effective in data-scarce situations like this and does not require any additional language-specific labeled training data. They exploited the linguistic similarities of the light-resource languages with any major language family to pre-training the zero-shot models. They used resources from any related resource-heavy language. A small example has been provided below.

Microsoft claims that during the online A/B testing of the spell-checker on Bing, no-result pages reduced by 30%, user-based manual intervention for query reform went below by 5%, spellings suggestions improved by 67%, and the click-through rate of the first page went up by a staggering 70%. The company seems to ramp-up the integration of Speller100 into its other services.

Advertisement

A Deep Dive Into IBM Quantum Roadmap

IBM Quantum Roadmap.jpg
Credit: IBM Quantum

“It took us 60 years from the first logic gates to modern cloud services. But IBM has set itself on a mission to fast forward the same journey for Quantum Computation (QC) to 3 years,” Jay Gambetta, IBM Fellow and Vice President, IBM Quantum.

Quantum Computing has opened up new doors to solve existing impenetrable problems and turn them into opportunities. But realizing the promised quantum power is still “The Road not taken.” To get the computations right, one needs exactly three things — the hardware stack, the software stack, and the developer environment. IBM has been a front-runner in the quantum race by settings above three things as early as 2010. And for each component, they had released a roadmap that is a brave move, considering that corporations usually post achievements rather than work plans. The company is bold enough to take chances and announced its ultimate goal openly — “to design a full-stack quantum computer deployed via the cloud that anyone around the world can program.”

IBM Quantum Hardware Roadmap

IBM wants to build scalable, larger, and better quantum computers, and the first objective, for now, is to create a 1000+ qubit system named Condor. But the researchers have to find out the solutions to preserve the states in the qubits for a more extended period while reducing noise-induced errors to make quantum computation viable. They have been busy optimizing the compilers, continuously refining  2-qubit gates, and many more to release the next candidate, ‘Eagle,’ surpassing 100+ qubit count and concurrently processing classical computations.

Also Read: IBM And Daimler Simulates Materials With Fewer Qubits

IBM Quantum Software Roadmap

The Circuit API is still in use for sending quantum instructions to quantum computers, which can handle smaller qubit systems. But to scale things up, more powerful circuits are needed that bring iterative phase estimation closer to the qubit systems than on the users’ system and cloud. 

IBM will soon release the new Qiskit runtime to boost the performance of the hybrid cloud technology 100 times. The Dynamic Circuits are also a force-multiplier that allows branching within the circuit based on measurements. These dynamic circuits can manipulate the future states based on the intermittent measurement and produce mid-circuit resets. This ability provides support for larger algorithmic complexities and circuit variety. As the software matures, some circuits will be used more frequently than the rest, so the goal is to offer a library of pre-built and optimized circuits for end-users. 

IBM Quantum Developer Environment Roadmap

IBM has shown that developers need not learn new tools or languages but instead use their existing code to interact with a quantum computer. All they need is a few lines of code to call a quantum API service on the cloud. The company is hopeful that the developers will be able to lay the foundation for the software stack that runs on the cloud accessible by anyone worldwide. In the long run, the developers and researchers from other fields shall seamlessly integrate quantum computing into their workflow.

Apart from the corporate cut-throating for profits from quantum services, the good thing that has happened is IBM’s support for the open-source developers and researchers building the quantum applications. The company is also busy outreaching to future generations by conducting wholesome courses, giving free credits to the quantum cloud, and providing ready-to-start developing materials like its Qiskit Handbook.

Advertisement

IIT Kanpur Offers Free 8-Weeks Computational Science Course, Enrollments Ends 15th Feb

IIT computational science course
Source - IIT Kanpur Gallery

IIT Kanpur has opened up the enrollment for an eight-week online course on computational science on the SWAYAM platform. An AvHumboldt Fellow with over 50 publications in his name, Dr. Ashoke De, a Professor in the Aerospace Engineering department, will teach the course.

The course is ideal for undergraduate and postgraduate students of Aerospace, Mechanical, Chemical, and Civil Engineering. However, a basic knowledge of mathematics and programming has been set as prerequisites. 

After taking this course, the learner will be able to leverage computation to solve various problems common to both pure and applied sciences. Learners can develop new methodologies and tools for carrying out numerical simulations, which is a significant part of the scientific computing paradigm. This course offers a basic overview of all these aspects that is easy to digest for beginners and faculties.

Also Read: AWS Will Host Free Virtual Classes On ML, Blockchain, Big Data, And More

When it comes to modeling natural phenomenons, knowledge of Linear Algebra, ODEs, and PDEs are a must. From the layout of the course, it is evident that the focus is on linear algebra for the first two weeks and Ordinary Differential Equations (ODEs) for the next two. Then, Partial Differential Equations (PDEs) are addressed in the fifth week.

On top of it, the course sharpens skills like mathematical modeling and numerical analysis. Therefore, the sixth-week module contains numerical analysis components focused on implementing eigenvalue problems and ODE solutions in the coming weeks. The course provides necessary insights into efficient algorithms, computer architecture, software design, implementation, validation, and results visualization.


This computational science course will be offered for free; hence, you can finish the whole course without paying a penny to the IIT. You have to pay a mere 1000 rupees for the exam to be conducted on 24th April for optional certification. You can join the course here before the enrollment deadline of 15th February.

Advertisement

Dealing With Racially-Biased Hate-Speech Detection Models

biased hate speech models

Hate-speech detection models are the most glaring example of biased models, as shown by researchers from Allen Institute for Artificial Intelligence in their linguistic study. In a recent post, the effects of statistical bias in machine translations were highlighted, but you shall see how dataset bias affects models in this post. The researchers studied the hate-speech detectors’ behavior using lexical — swear words, slurs, identity mentions — and dialectal markers — specifically African-American English. They also proposed an automated dialect-aware data correction method, which uses synthetic labels to reduce dialectal associations with toxicity score.

The dataset creation process always captures biases that are inherent to humans. This dataset bias consists of spurious correlations between surface patterns and annotated toxicity labels. These spurious correlations give rise to two different types of bias, lexical and dialectical. The lexical bias associates toxicity with certain words that are considered profane and identity mentions, while dialectal bias correlates toxicity with the lingua franca of minorities. All these biases proliferate freely during the training phase of the hate-speech models. 

Researchers have proposed numerous debiasing techniques in the past, some applied by internet giants — Google, Facebook, and Twitter — in their systems. In this study, the researchers found that these models are not good enough. The so-called “Debiased” models still disproportionately flag text in particular dialects as toxic. The researchers noted, ”mitigating dialectal bias through current debiasing methods does not mitigate a model’s propensity to label tweets by black authors as more toxic than by white authors.”

A proof-of-concept solution was proposed by the Allen researchers that ward off the problem. The idea is to parse those “reported” hate-speeches into the majority’s lingua franca deemed non-toxic by the classifier. This idea takes care of the speeches’ dialectal context, resulting in common ground for the model to predict the toxicity score of speeches reasonably and be less prone to dialectal and racial biases.

Advertisement

AutoML Made Easy With Symbolic Programming using Pyglove

Symbolic programing AutoML Pyglove

Google AI researchers have released a PyGlove library, a symbolic implementation of Automated Machine Learning (AutoML) that allows developers to experiment with search spaces, search algorithms, and search flows of an AutoML with only a few code lines. Now, developers can self-mutate Python classes and functions through brief Python annotations, making it much easier to write AutoML programs.

Developers previously had data and the outputs; they fed that into a machine learning algorithm, which automated the learning of rules governing input to output. Researchers later automated the selection and hyper-parameter tuning of those machine learning algorithms as well. One of the sub-classes of machine learning algorithms is neural networks, which are highly sensitive to architecture and hyper-parameters.

The possible combinations of architecture and hyper-parameter choices become humongous as researchers aim to build larger and larger neural models. They waste months in hand-crafting neural network architectures and selecting the right hyper-parameters. AutoML automated these aspects by formulating the problem as a search problem.

Also Read: What Is Liquid Machine Learning?

A search space is defined to represent all possible choices, and a search algorithm is used to find the best options. Neural Architecture Search (NAS) algorithms like ENAS and DARTS come under the purview of the AutoML. But the current implementations of NAS algorithms do not offer modularity to the components of NAS algorithms like the search space and search algorithm. Therefore, researchers had to face difficulties modifying the search space, search algorithm, or search flow alone.

The Google researchers introduced AutoML based on symbolic programming — a paradigm that allows self-mutating programs by manipulating its components — that makes components decoupled. This decoupling makes it easy for practitioners to change the search space and search algorithm with and without weight sharing, and add search capabilities to existing code and implement complex search flows.

On ImageNet and NAS-Bench-101 benchmarks, they showed that symbolic programming based PyGlove converts a static program into a search space, quickly iterates on the search spaces and search algorithms, and crafts complex search flows to achieve better results. PyGlove also allows easy plug-and-play of AutoML techniques in existing ML pipelines while also benefiting open-ended research.

Advertisement

Microsoft Introduces VIVA To Help People Work From Home

Microsoft VIVA employee platform

Microsoft has recently unveiled a new employee experience platform–VIVA–that will act as an integrated platform to manage employee well-being, learning, engagement, and knowledge discovery in the workflow. With close integration of Teams and Office 365 technologies, Microsoft wants to be the market leader in employee engagement.

Microsoft has bet upon the remote working culture in the future. It is targeting all kinds of organizations and employees. During the pandemic, almost all companies have been distraught with various platforms for on-boarding and training the employees. This new platform promises to smoothen the journey for both employees and the company alike. 

Also Read: AWS Will Host Free Virtual Classes On ML, Blockchain, Big Data, And More

Currently, the platform has four foundations — Viva Connect, Viva insight, Viva learning, and Viva topics — each one represents a different aspect of employee workflow from in-vitro the company or outside. Viva Connect has a personalized gateway for every employee to access all internal communications and company resources. It also helps employees participate in communities like employee resource groups, all from a single customizable app in Microsoft Teams. VIVA Insight helps the CXOs identify where teams struggle, especially to balance productivity and well-being.

VIVA Learning makes learning resources available in the company, like courses and guided-projects from EDX, Coursera, and many more, in one platform. It helps the employees to manage all their training and micro-courses with their accomplishments. VIVA Topics allows knowledge discovery from various third-party sources inside the documents across Microsoft 365 and conversation in Teams using AI.

Microsoft has partnered with Accenture, Avanade, PwC, and EY to help other companies adopt the homegrown employee experience environment, providing consulting and advisory services. Microsoft’s CEO Satya Nadela put the benefits of VIVA into a much-needed vision statement amidst this new initiative. He said, “We have participated in the largest at-scale remote work experiment the world has seen, and it has had a dramatic impact on the employee experience. Every organization will require a unified employee experience from onboarding and collaboration to continuous learning and growth. Viva brings together everything an employee needs to be successful, from day one, in a single, integrated experience directly in Teams.”

Advertisement

AWS Will Host Free Virtual Classes On ML, Blockchain, Big Data, And More

AWS virtual classes

AWS will offer free virtual classes for learners who want to gain in-demand skills like machine learning, Blockchain, big data, containers, and more. The AWS virtual class is an ideal way to getting started with the latest technologies on AWS. 

The webinar-based online classes are about 90 minutes that are mostly focused on beginners or professionals who want to explore new technologies.

Learners from across the world can register for the event based on their convenience as the lessons are delivered across the timezone. AWS keeps hosting training through virtual classes to ensure the world has a workforce that can work with cutting-edge technologies. Last month, AWS conducted free AWS AI Conclave to deliver 20+ breakthrough sessions from the industry experts.

In February 2021, AWS included topics like BlockchainContainersKubernetes, Machine LearningBig Data, among others. With these webinars, aspirants can find new interests and get to know the fundamentals of technologies.

As organizations are expecting fundamental knowledge of product development from data scientists, gaining an understanding of containers can differentiate them among others. In addition to learning new technologies, familiarising with the AWS platform can also help beginners in streamlining the workflow while they start working at organizations.

Every month, AWS delivers webinars on the latest technologies, and in the coming months, AWS will also focus on Cloud Security, Data Analytics, and more.

To know more about the upcoming AWS Virtual Classes click here.

Advertisement

Measuring Weirdness In AI-Based Language-Translations

machine translation linguistic analysis

AI-based language translations were the object of ridicule when they coughed up something funny. Consequently, AI researchers focused on translation accuracy and preserved their fluidness to set aside the embarrassment because of faulty translations. The situation gradually improved, especially with better and larger language models that surpassed humans in various benchmarks.

But these language models still amplify the statistical biases found in their training data. And the biases affect not only the translations but also their linguistic richness. Researchers from the University of Maryland and Tilburg University have tried to study this effect quantitatively in terms of grammar and linguistic analysis of machine translations. 

The translated work differs from the original one thanks to intentional factors like explicitation and normalization and unintentional ones like unconscious effects of the source language input on the target language produced. These factors are studied under a linguistics field, called Translationese, to assess the translator’s unique additions. Similarly, linguists analyze these elements introduced by a machine translator under Machine Translationese.

Also Read: Language Models Exhibits Larger Social Bias Than Human-Written Texts

In the study, the researcher linguistic analysis of sequential neural models like LSTMs, Transformers, and phrase-based statistical translation models to highlight the above factors. These models were tasked with translation between English, French, and Spanish from the source. They found that the statistical distribution of terms in the training data dictates the morphological loss of variety in the machine translations.

The translation systems do not distinguish between the synonymous and grammatical variants. This directly reduces the number of grammatically correct but diverse options. In layman terms, the diversity of words and sentence structure was drastically low in the translations because of consistency and simplification.

The authors also investigated the impacts of the loss in social-lingual aspects because these machine translations affect language usage among the masses. No solution has been proposed to the problem. The authors believe that different metrics like language acquisition metric to analyze lexical sophistication, Shannon entropy, and Simpson diversity to study morphological diversity, shall contribute further investigation.

Advertisement