Home
Services
Services
Develop your AI maturity with tailored services
01
AI & data strategy
02
Data engineering
03
Educational workshop
04
Tailored ML solutions
05
AI team extension
06
ML deployment & operations
Industries
01
Chemicals
02
CPG
03
Banking
04
Logistics
05
Manufacturing
06
Pharma
07
Utilities
Industries
Industries
Ensure the success of your AI transformation
01
Chemical
02
Consumer Packaged Goods
03
Financial Services
04
Logistics
05
Manufacturing
06
Pharmaceuticals
07
Utilities
About
Our valuesOur story
Our team
Blog
All postsAI & MLAbout VisiumCase studiesDigestTalks & Webinars
JOIN USCONTACT US
Posted on 
February 24, 2021
|
minutes

Material Generation, AI for Customer Support, and Stochastic Parrots

Using LSTMs for efficient material generation

Research from Sandia National Laboratories accelerates phase-field micro-structure evolution predictions using surrogate Machine Learning methods

Context

When creating new materials, it is extremely important to know how they will react for a wide variety of physical, chemical, and biological conditions. To grasp the evolution of a micro-structure's properties through these varying conditions, the phase-field method is used. It is a powerful and versatile computational approach for modeling the material's behavior. Unfortunately, high-fidelity phase-field models are computationally expensive. To require an actionable accuracy, one needs high-performance computing resources and sophisticated integration schemes.

What's new

In a recent publication, researchers from Sandia National Laboratories leveraged data from past phase-field simulations to predict the effect of a change in micro-structure design to the material's properties.

They started with dimensionality reduction using Principal Component Analysis, which enables the construction of a low-dimensional but faithful representation of the time evolution of the micro-structure. The researchers use a surrogate machine-learned model that is trained with an LSTM architecture. Using high-fidelity data from 300,000 time observations (5,000 simulations of 60 frames), they were able to achieve an impressive ARE (Absolute Relative Error) of 6.8%.

Their LSTM model is able to simulate thermodynamic processes (e.g. congealing of a cooling molten alloy) in only 60 ms compared to the usual 12 minutes. "As such, the computational efficiency of the LSTM model yields results 42,666 times faster than the full-scale phase-field method."

a Reconstructed microstructure from the LSTM-trained surrogate model using a phase-recovery algorithm. b Phase-field predictions using LSTM-trained surrogate model as an input. c Point-wise error between predicted and true microstructure evolution. d Cumulative probability distribution of the absolute relative error on characteristic microstructural feature size. e Comparison of radial average of the microstructure autocorrelation between predicted (red) and true (black) microstructure evolution.

Why it matters

In recent years, the use of Machine Learning models has shown incredible results in augmenting scientific simulations and acquisition techniques. Most notably, DeepMin's AlphaFold solves the protein folding problem and DENSE, from researchers at Stanford, builds high accuracy emulators.

Many impressive advances are also seen in less publicized domains. For instance, Martin Weigert and colleagues leverage Machine Learning to push the limits imposed on fluorescence microscopy by optics, fluorophore chemistry, and maximum photon exposure allowed by living samples.

What's next

The pace of scientific discovery is ever-increasing. The application of Machine Learning to augment computational techniques as well as acquisition systems yields great promise for the fields of aerospace, energy, medicine, and others. Advanced materials are changing the world, and these new techniques are helping these innovations reach the market faster than ever before.

‍

What can AI do for customer support?

AI is transforming customer service interactions, ensuring that all angles are covered in terms of customer support

Context

In recent years, customer service has taken center-stage. Shopping has become extremely customer-centric, providing shoppers with highly targeted and tailored experiences. Along with this shift, all eyes are directed at the biggest predictor of churn and customer lifetime-value: customer service. By transforming customer service interactions, tailored experiences are prepared to improve every aspect of your business including online customer experience, loyalty, brand reputation, preventive assistance, and even generation of revenue streams.

Source: Zendesk survey

What's new

Customer service is going digital. There's no doubt about it. A Zendesk survey shows that excellent customer service "directly impacts long-term revenue" and "requires a wide range of channels". How can an organization ensure wide availability and quick responses at the same time? The answer is quite simple: the implementation of data-driven and automated solutions.

  • Leveraging AI as a brand messenger

By developing a chat-bot that responds to customer input (text, selection, voice, etc.), companies are able to inform their customers and potential customers seamlessly. The bot is available 24/7 and helps customers easily obtain the information they need. This solution is gaining a lot of traction in industries such as fashion, tourism, food chains, e-commerce, as well as hotels and airlines.

Source: KLM
  • Ticket sorting

Implement an intelligent sorting algorithm to prioritize ticket queues for your customer service representatives. The Machine Learning algorithms allow organizations to use the tickets' features as inputs and yield an optimal order of resolution. The main benefit of such a solution is that a company can improve the quality and efficiency of its customer service by prioritizing important issues.

  • Customer lifetime value prediction

While more of a back-office task, the prediction of customer lifetime value (CLV) has immense value. Using historical transaction information from retail and e-commerce stores, CLV is predicted for customers, old and new. The insights are immense they can help organisations guide their assortment and offering of products as well as marketing and promotional campaigns, which have an immense influence on revenue.

  • Using Machine Learning for extra support

So much time is lost looking for information in an organisation. Long chains of emails, endless Google searches, and calls to diverse departments from multiple offices in varying time-zones, you name it! When time is essential, as it is in customer service, this becomes a business hurdle. By centralizing and giving structure to internal documents, Machine Learning algorithms are capable of searching through it and flagging the most relevant content. An internal Google? Not quite. But the data never lies, intelligent knowledge search for private data is the key to increasing the efficiency of customer service teams.

Why it matters

With the increasing demand for informative and timely customer service and its threat of churn, companies across the globe are starting to implement digital solutions. Whether these systems interact with the client or not, they leverage historical and public data sources to increase the efficiency of their customer support.

What's next

As AI helpers such as chat-bots and smart search platforms become the new norm, organizations are faced with a digital transformation challenge. Visium's expertise in providing customer support solutions in multiple and diverse formats allows organizations in Switzerland and abroad to tackle this challenge, head first.

‍

The dangers of Large Language Models

Authors have raised global awareness on recent NLP trends and have urged researchers, developers, and practitioners associated with language technology to take a holistic and responsible approach.

Context

Large Language Models (LLMs) are NLP models that leverage large datasets with a huge amount of parameters on string prediction tasks. This means they predict the likelihood of a token (which can be a character, word, or string) given the previous or surrounding tokens. The goal is to train the model in an unsupervised manner initially and then fine-tune it for specific tasks. Most commonly, you will see them out in the wild outputting scores or string predictions when given text as input. Well-known examples are BERT and its derivatives, T-NLG, and GPT-3.

What's new

In recent years, LLMs have increased in size: meaning they are trained with larger amounts of data and parameters. In fact, the Transformer architecture yields continuously noticeable benefits from this size increase. For this reason, we can expect size of LLMs to increase as long as this correlates with better performance.

Source: Bender et al. 2021

Evidently, the training of models on such large scales implies risks and costs that are not attributed to smaller models. These dangers are explored in a recent paper by researchers from the University of Washington and Black in AI and can be summarized as:

  • Environmental cost: Training a BERT base model without fine-tuning is estimated to require as much energy as a trans-American flight (~1900 CO2e).
  • Financial cost: in the past 6 years, the compute required to train the largest Deep Learning models has increased 300'000 fold. This raises questions about fairness, as only the world's top AI research institutions are capable of supporting such financial costs. Furthermore, most of these models are trained on the English language, and their use is not geared to those who could most benefit from it (e.g. marginalized communities).
  • Training with large data: the data used for these models are often collected from the internet. As internet access and contributions are not evenly distributed, this poses a risk for the over-representation of younger people in more developed countries. Above that, the datasets are undocumented and too large to document post hoc. While documentation allows for potential accountability, undocumented training data perpetuates harm without recourse.
  • Dangers of deploying LLMs at scale: human communication relies greatly on the implicit meaning that's conveyed with language. Text produced by LLMs will reproduce and even amplify the encoded biases in their training data.
Source: Bender et al. 2021

For more details, click here to access the original paper.

Why it matters

The world's top researchers are putting in immense time and effort to work on LLMs.

As stated by the authors: "In order to mitigate the risks that come with the creation of increasingly large LMs, we urge researchers to shift to a mindset of careful planning, along many dimensions, before starting to build either datasets or systems trained on datasets. We should consider our research time and effort a valuable resource, to be spent to the extent possible on research projects that build towards a technological ecosystem whose benefits are at least evenly distributed or better accrue to those historically most marginalized."

What's next

Currently, methods are being developed to counter these costs and risks. For instance, a recent paper was published on increasing the efficiency of Transformers with Performers.

In general, the authors of "On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?" give six guidelines for future research:

  1. Considering Environmental and Financial Impacts.
  2. Doing careful data curation and documentation.
  3. Engaging with stakeholders early in the design process.
  4. Exploring multiple possible paths towards long-term goals.
  5. Keeping alert to dual-use scenarios.
  6. Allocating research effort to mitigate harm.

‍

Tagged:
NLP
Retail
Manufacturing
view All Posts
Subscribe
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Featured Posts
Digest
Data Cascades, AI for Football, and Protein Generation
Digest
AI in Manufacturing, Google Vertex AI, and Session-Based Recommendations
Digest
Green ML, Interpretable Cancer Detection, and Self-supervised Transformers
Digest
AI Regulations, Car Wreckognition, and External Data Copy
Digest
Melanoma Detection, Bank Customer Confidence and Welding Control
Tags
AI For Good
AI Governance
Document Management
Ethics in AI
Healthcare
Keynotes and Talks
ML in Production
Manufacturing
NLP
Probabilistic Programming
R&D
Retail
Sound AI
Stay Connected
Subscribe

Get new posts to your inbox

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
More Posts

You might also like

AI & ML
AI in Financial Services: 10 questions you should ask before thinking about leveraging new technologies for AML- Fraud detection
Nov 8, 2021
 by 
Digest
Data Cascades, AI for Football, and Protein Generation
Jul 26, 2021
 by 
Digest
AI in Manufacturing, Google Vertex AI, and Session-Based Recommendations
Jun 15, 2021
 by 
AI & ML
Let’s Talk AI Strategy
Jun 4, 2021
 by 
Lucas Nottaris
Digest
Green ML, Interpretable Cancer Detection, and Self-supervised Transformers
Jun 1, 2021
 by 
Digest
AI Regulations, Car Wreckognition, and External Data Copy
May 18, 2021
 by 

We tailor state-of-the-art AI solutions for the world's best brands

Navigation
HomeAboutBlogDigestContactPrivacy policy
Services
AI & data strategyIdeation workshopTailored ML solutionsAI team extensionML operations
© Copyright 2022. All Rights Reserved.
Visium SA