Rachit Ranjan, Author at MarkTechPost https://www.marktechpost.com/author/rachit-ranjan/ An Artificial Intelligence News Platform Wed, 25 Oct 2023 00:04:26 +0000 en-US hourly 1 https://wordpress.org/?v=6.3.2 https://www.marktechpost.com/wp-content/uploads/2022/04/cropped-Favicon-512-x-512-1-1-32x32.png Rachit Ranjan, Author at MarkTechPost https://www.marktechpost.com/author/rachit-ranjan/ 32 32 127842392 Google DeepMind Proposes An Artificial Intelligence Framework for Social and Ethical AI Risk Assessment https://www.marktechpost.com/2023/10/25/google-deepmind-proposes-an-artificial-intelligence-framework-for-social-and-ethical-ai-risk-assessment/ https://www.marktechpost.com/2023/10/25/google-deepmind-proposes-an-artificial-intelligence-framework-for-social-and-ethical-ai-risk-assessment/#respond Wed, 25 Oct 2023 11:00:00 +0000 https://www.marktechpost.com/?p=45094 Generative AI systems, which create content across different formats, are becoming more widespread. These systems are used in various fields, including medicine, news, politics, and social interaction, providing companionship. Using natural language output, these systems produced information in a single format, such as text or graphics. To make generative AI systems more adaptable, there is […]

The post Google DeepMind Proposes An Artificial Intelligence Framework for Social and Ethical AI Risk Assessment appeared first on MarkTechPost.

]]>

Generative AI systems, which create content across different formats, are becoming more widespread. These systems are used in various fields, including medicine, news, politics, and social interaction, providing companionship. Using natural language output, these systems produced information in a single format, such as text or graphics. To make generative AI systems more adaptable, there is an increasing trend in improving them to operate with additional formats, such as audio (including voice and music) and video.

The increasing use of generative AI systems highlights the need to assess potential risks associated with their deployment. As these technologies become more prevalent and integrated into various applications, concerns arise regarding public safety. Consequently, evaluating the potential risks posed by generative AI systems is becoming a priority for AI developers, policymakers, regulators, and civil society.

The growing use of these systems highlights the necessity to evaluate potential dangers related to implementing generative AI systems. As a result, it is becoming more important for AI developers, regulators, and civil society to assess the possible threats posed by generative AI systems. The development of AI that might spread false information raises moral questions about how such technologies will affect society.

Consequently, a recent study by Google DeepMind researchers offers a thorough approach to assessing AI systems’ social and ethical hazards across several contextual layers. The DeepMind framework systematically assesses risks at three distinct levels: the system’s capabilities, human interactions with the technology, and the broader systemic impacts it may have. 

They emphasized that it is crucial to recognize that even highly capable systems may only necessarily cause harm if used problematically within a specific context. Also, the framework examines real-world human interactions with the AI system. This involves considering factors such as who utilizes the technology and whether it operates as intended.

Finally, the framework checks how AI delves into the risks that may arise when AI is extensively adopted. This evaluation considers how technology influences larger social systems and institutions. The researchers emphasize how important context is in determining how risky AI is. Each layer of the framework is permeated by contextual concerns, emphasizing the importance of knowing who will use the AI and why. For instance, even if an AI system produces factually accurate outputs, users’ interpretation and subsequent dissemination of these outputs may have unintended consequences only apparent within certain contextual constraints.

The researchers provided a case study concentrating on misinformation to demonstrate this strategy. The evaluation includes assessing an AI’s tendency for factual errors, observing how users interact with the system, and measuring any subsequent repercussions, such as the spread of incorrect information. This interconnection of model behavior with actual harm occurring in a given context leads to actionable insights.

DeepMind’s context-based approach underscores the importance of moving beyond isolated model metrics. It emphasizes the critical need to evaluate how AI systems operate within the complex reality of social contexts. This holistic assessment is crucial for harnessing the benefits of AI while minimizing associated risks.


Check out the Paper. All Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post Google DeepMind Proposes An Artificial Intelligence Framework for Social and Ethical AI Risk Assessment appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/25/google-deepmind-proposes-an-artificial-intelligence-framework-for-social-and-ethical-ai-risk-assessment/feed/ 0 45094
Meet MatFormer: A Universal Nested Transformer Architecture for Flexible Model Deployment Across Platforms https://www.marktechpost.com/2023/10/21/meet-matformer-a-universal-nested-transformer-architecture-for-flexible-model-deployment-across-platforms/ https://www.marktechpost.com/2023/10/21/meet-matformer-a-universal-nested-transformer-architecture-for-flexible-model-deployment-across-platforms/#respond Sat, 21 Oct 2023 16:27:15 +0000 https://www.marktechpost.com/?p=44930 Transformer models find applications in various applications, ranging from powerful multi-accelerator clusters to individual mobile devices. The varied requirements for inference in these settings make developers train fundamental models like PaLM 2, Llama, and ViTs in different sizes. However, the higher costs associated with training lead to a restricted set of supported model sizes.  Large […]

The post Meet MatFormer: A Universal Nested Transformer Architecture for Flexible Model Deployment Across Platforms appeared first on MarkTechPost.

]]>

Transformer models find applications in various applications, ranging from powerful multi-accelerator clusters to individual mobile devices. The varied requirements for inference in these settings make developers train fundamental models like PaLM 2, Llama, and ViTs in different sizes. However, the higher costs associated with training lead to a restricted set of supported model sizes. 

Large foundational models are used in different situations, such as giving quick responses on mobile phones or handling batches on multi-cluster GPUs for large-scale web applications. Each model provides a selection of independently trained models in different sizes to accommodate various circumstances. To accommodate a wide range of applications, these model sizes are typically grouped on a logarithmic scale in a roughly linear fashion.

Consequently, a group of researchers from Google Research, the University of Texas at Austin, the University of Washington, and Harvard University have introduced MatFormer—a Transformer architecture explicitly crafted for adaptability, as outlined in their latest paper, which is titled MatFormer: Nested Transformer for Elastic Inference. MatFormer makes it easier to build an integrated model that can generate numerous smaller submodels without extra training.

They have incorporated a nested sub-structure within the standard Transformer and jointly optimized all the granularities to produce a single, universal elastic model.

The researchers emphasized that they have produced many accurate submodels without acquiring additional training costs by deliberately mixing various levels of information in various layers of a universal MatFormer model. Each Feed Forward Network (FFN) block in the MatFormer architecture is optimized with a collection of smaller, nested FFN blocks. Each Feed Forward Network (FFN) block in the MatFormer architecture is optimized with a collection of smaller, nested FFN blocks. Through this training approach, they combined and adjusted the complexity of the model across different layers. 

The nested structure is implemented on the hidden representations of the Feed Forward Network (FFN) block, amplifying the model’s capabilities by placing the attention heads in order of significance. A substructure within the attention heads is created from the most to the least. Compared to independently training equivalent Transformer-based submodels, training is accelerated by 15% since the more significant heads are distributed among a larger number of submodels. Additionally, this method aligns with the specifically optimized submodel curve and permits the extraction of several smaller submodels while maintaining accuracy.

The researchers found that they could produce a sizable number of accurate smaller models without further optimization by choosing different levels of detail for each MatFormer layer.

The team studied the effectiveness across a range of model types (decoders and encoders), modalities (language and vision), and scales (up to 2.6 billion parameters). The researchers emphasized that comparing these smaller models to their independently trained counterparts reveals comparable validation loss and one-shot downstream performance. Also, MatFormer exhibits robust generalization and works well as vision encoders (MatViT) and decoder-only language models (MatLM). In terms of accuracy and dependability, it scales similarly to the traditional Transformer. 


Check out the PaperAll Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post Meet MatFormer: A Universal Nested Transformer Architecture for Flexible Model Deployment Across Platforms appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/21/meet-matformer-a-universal-nested-transformer-architecture-for-flexible-model-deployment-across-platforms/feed/ 0 44930
6 Magic Commands for Jupyter Notebooks in Python Data Science https://www.marktechpost.com/2023/10/20/6-magic-commands-for-jupyter-notebooks-in-python-data-science/ https://www.marktechpost.com/2023/10/20/6-magic-commands-for-jupyter-notebooks-in-python-data-science/#respond Fri, 20 Oct 2023 18:42:14 +0000 https://www.marktechpost.com/?p=44879 In the field of Python-based Data Science projects, the utilization of Jupyter Notebooks is ubiquitous. These interactive and user-friendly environments facilitate seamless integration of code and documentation, providing a conducive space for exploration and analysis. Within this framework exists a set of magic commands that prove invaluable tools. These commands enhance workflow efficiency and serve […]

The post 6 Magic Commands for Jupyter Notebooks in Python Data Science appeared first on MarkTechPost.

]]>

In the field of Python-based Data Science projects, the utilization of Jupyter Notebooks is ubiquitous. These interactive and user-friendly environments facilitate seamless integration of code and documentation, providing a conducive space for exploration and analysis. Within this framework exists a set of magic commands that prove invaluable tools. These commands enhance workflow efficiency and serve as time-saving instruments for the discerning data scientist.

1. Conversing with Models in Jupyter

The command “%%ai” makes it possible to enter the world of natural language interactions with machine learning models. Users can choose a model using this command and then have natural language conversations with that model. This function expands the range of possibilities for model exploration and enhances the interactivity of Jupyter Notebooks.

 2.%%latex: Elevating Visual Representations

The “%%latex” command must include mathematical equations or symbols in their notebooks. The rendering of LaTeX code directly in Jupyter Notebooks is made possible by this command, providing the seamless integration of mathematical expressions for clearer and more expert presentations.

3. %%sql: Empowering Database Interactions

With the “%%sql” magic instructions, the integration of SQL queries into Jupyter Notebooks is simplified. It allows the users to execute SQL queries directly inside the notebook environment. This functionality eliminates the need for external interfaces, which is useful for data scientists using databases.

4. %run: Effortless Python File Execution

With the “%run” magic command, running external Python files inside a Jupyter Notebook is simpler. Only one command is needed to access the data inside a Python file, whether a standalone script or module. This improves Jupyter-based applications’ modularity by making it easier to integrate external code easily.

5. %%writefile: Streamlining File Creation

The magic command “%%writefile” takes care of the necessity for quick file creation within the notebook. Users can easily create new Python files by entering the desired file name and including the content within the cell. This functionality guarantees a simpler approach to file management while improving code organization.

 6. %history -n: Retrieving Previous Commands

In Jupyter Notebooks, sometimes we accidentally delete our commands and the results they give. But there’s a helpful trick called “%history -n.” With this, we can see a list of all your past commands, and you can decide how many of them you want to look at (“-n” lets you choose).

For Python-based Data Science projects, the integration of these magic commands enhances the Jupyter Notebook experience. These commands greatly improve workflow through interactions with models, effective database interactions, and simplified file management. Having such tools becomes crucial for remaining ahead in searching for insights and discoveries as the data science landscape changes. Using these commands, data scientists can make their projects less complicated and work better. This will make their studies stronger and more important in the end.


Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post 6 Magic Commands for Jupyter Notebooks in Python Data Science appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/20/6-magic-commands-for-jupyter-notebooks-in-python-data-science/feed/ 0 44879
Microsoft Researchers Propose DeepSpeed-VisualChat: A Leap Forward in Scalable Multi-Modal Language Model Training https://www.marktechpost.com/2023/10/20/microsoft-researchers-propose-deepspeed-visualchat-a-leap-forward-in-scalable-multi-modal-language-model-training/ https://www.marktechpost.com/2023/10/20/microsoft-researchers-propose-deepspeed-visualchat-a-leap-forward-in-scalable-multi-modal-language-model-training/#respond Fri, 20 Oct 2023 09:29:54 +0000 https://www.marktechpost.com/?p=44874 Large language models are sophisticated artificial intelligence systems created to understand and produce language similar to humans on a large scale. These models are useful in various applications, such as question-answering, content generation, and interactive dialogues. Their usefulness comes from a long learning process where they analyze and understand massive amounts of online data. These […]

The post Microsoft Researchers Propose DeepSpeed-VisualChat: A Leap Forward in Scalable Multi-Modal Language Model Training appeared first on MarkTechPost.

]]>

Large language models are sophisticated artificial intelligence systems created to understand and produce language similar to humans on a large scale. These models are useful in various applications, such as question-answering, content generation, and interactive dialogues. Their usefulness comes from a long learning process where they analyze and understand massive amounts of online data.

These models are advanced instruments that improve human-computer interaction by encouraging a more sophisticated and effective use of language in various contexts.

Beyond reading and writing text, research is being carried out to teach them how to comprehend and use various forms of information, such as sounds and images. The advancement in multi-modal capabilities is highly fascinating and holds great promise. Contemporary large language models (LLMs), such as GPT, have shown exceptional performance across a range of text-related tasks. These models become very good at different interactive tasks by using extra training methods like supervised fine-tuning or reinforcement learning with human guidance. To reach the level of expertise seen in human specialists, especially in challenges involving coding, quantitative thinking, mathematical reasoning, and engaging in conversations like AI chatbots, it is essential to refine the models through these training techniques.

It is getting closer to allowing these models to understand and create material in various formats, including images, sounds, and videos. Methods, including feature alignment and model modification, are applied. Large vision and language models (LVLMs) are one of these initiatives. However, because of problems with training and data availability, current models have difficulty addressing complicated scenarios, such as multi-image multi-round dialogues, and they are constrained in terms of adaptability and scalability in various interaction contexts.

The researchers of Microsoft have dubbed DeepSpeed-VisualChat. This framework enhances LLMs by incorporating multi-modal capabilities, demonstrating outstanding scalability even with a language model size of 70 billion parameters. This was formulated to facilitate dynamic chats with multi-round and multi-picture dialogues, seamlessly fusing text and image inputs. To increase the adaptability and responsiveness of multi-modal models, the framework uses Multi-Modal Causal Attention (MMCA), a method that separately estimates attention weights across several modalities. The team has used data blending approaches to overcome issues with the available datasets, resulting in a rich and varied training environment.

DeepSpeed-VisualChat is distinguished by its outstanding scalability, which was made possible by thoughtfully integrating the DeepSpeed framework. This framework exhibits exceptional scalability and pushes the limits of what is possible in multi-modal dialogue systems by utilizing a 2 billion parameter visual encoder and a 70 billion parameter language decoder from LLaMA-2. 

The researchers emphasize that DeepSpeed-VisualChat’s architecture is based on MiniGPT4. In this structure, an image is encoded using a pre-trained vision encoder and then aligned with the output of the text embedding layer’s hidden dimension using a linear layer. These inputs are fed into language models like LLaMA2, supported by the ground-breaking Multi-Modal Causal Attention (MMCA) mechanism. It is significant that during this procedure, both the language model and the vision encoder stay frozen.

According to the researchers, classic Cross Attention (CrA) provides new dimensions and problems, but Multi-Modal Causal Attention (MMCA) takes a different approach. For text and image tokens, MMCA uses separate attention weight matrices such that visual tokens focus on themselves and text permits focus on the tokens that came before them.

DeepSpeed-VisualChat is more scalable than previous models, according to real-world outcomes. It enhances adaption in various interaction scenarios without increasing complexity or training costs. With scaling up to a language model size of 70 billion parameters, it delivers particularly excellent scalability. This achievement provides a strong foundation for continued advancement in multi-modal language models and constitutes a significant step forward.


Check out the Paper and GithubAll Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post Microsoft Researchers Propose DeepSpeed-VisualChat: A Leap Forward in Scalable Multi-Modal Language Model Training appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/20/microsoft-researchers-propose-deepspeed-visualchat-a-leap-forward-in-scalable-multi-modal-language-model-training/feed/ 0 44874
Revolutionizing Wearable Tech: Edge Impulse’s Ultra-Efficient Heart Rate Algorithm & Expanding Healthcare Suite https://www.marktechpost.com/2023/10/17/revolutionizing-wearable-tech-edge-impulses-ultra-efficient-heart-rate-algorithm-expanding-healthcare-suite/ https://www.marktechpost.com/2023/10/17/revolutionizing-wearable-tech-edge-impulses-ultra-efficient-heart-rate-algorithm-expanding-healthcare-suite/#respond Tue, 17 Oct 2023 09:09:26 +0000 https://www.marktechpost.com/?p=44703 Machine learning is used in almost every aspect of our lives and across various fields. It’s a technology becoming increasingly prevalent and finding applications in many areas. Its relevance is especially important in medicine because it is essential to improving healthcare procedures. Machine learning is revolutionizing how we tackle medical problems, from identifying diseases to […]

The post Revolutionizing Wearable Tech: Edge Impulse’s Ultra-Efficient Heart Rate Algorithm & Expanding Healthcare Suite appeared first on MarkTechPost.

]]>

Machine learning is used in almost every aspect of our lives and across various fields. It’s a technology becoming increasingly prevalent and finding applications in many areas. Its relevance is especially important in medicine because it is essential to improving healthcare procedures. Machine learning is revolutionizing how we tackle medical problems, from identifying diseases to forecasting patient outcomes, ultimately leading to better patient care and medical research.

Consequently, a company called Edge Impulse, which specializes in on-device machine learning and artificial intelligence, has announced the launch of what it claims is the smallest and most precise heart rate measurement algorithm. They also emphasized that it requires only one-sixteenth of the competition’s memory.

The researchers emphasize that this innovative algorithm functions as a nervous system health detective for our body. To comprehend how the autonomic nervous system is maintained in balance, it examines changes in our heart rate and the intervals between beats. Our general health, including heart health, stress levels, and how quickly we bounce back from activities, depends on this balance. 

With the help of a straightforward sensor that measures the light that passes through our skin (a photoplethysmogram), the algorithm’s cleverness allows it to provide precise heart rate and heart rate variability values. Wearables like those worn on the finger frequently contain this sensor. The measurement and analysis of heart rate interbeat intervals (IBIs) are fundamental in studying cardiovascular physiology and health. Heart rate variability (HRV) measures the variation in time between successive heartbeats. It goes beyond the measurement of the heart rate itself. 

The algorithm primarily uses light-based sensors like those used in fitness bands and smartwatches, but it can also utilize electrocardiogram (ECG) sensors. It is extremely intelligent—while using only one-sixteenth of the memory compared to its nearest rival, it can diagnose atrial fibrillation, detect falls, monitor sleep, gauge stress, and recognize changes in activity levels.

They have algorithms for measuring body temperature, monitoring movement, and tracking posture and brain activity data through electroencephalograms (EEG). Edge Impulse has developed data dashboards for real-time monitoring and a research data lake for clinical data to improve these algorithms even more.

The researchers emphasized that this significantly reduces the money required for research and development (R&D) to produce unique algorithms. The researchers also highlighted that modern algorithms are used in Edge Impulse’s HR/HRV solutions, which negates the need for time-consuming, difficult algorithm refinement. 

Edge Impulse also offers a robust infrastructure to enable the growth of centralized and decentralized clinical investigations, accommodating small and big subject groups. This scalability is essential for extensive testing and validation since it guarantees that the dataset utilized is diverse and reduces model biases.


Check out the Research Page and Reference ArticleAll Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post Revolutionizing Wearable Tech: Edge Impulse’s Ultra-Efficient Heart Rate Algorithm & Expanding Healthcare Suite appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/17/revolutionizing-wearable-tech-edge-impulses-ultra-efficient-heart-rate-algorithm-expanding-healthcare-suite/feed/ 0 44703
Apple and CMU Researchers Unveil the Never-ending UI Learner: Revolutionizing App Accessibility Through Continuous Machine Learning https://www.marktechpost.com/2023/10/14/apple-and-cmu-researchers-unveil-the-never-ending-ui-learner-revolutionizing-app-accessibility-through-continuous-machine-learning/ https://www.marktechpost.com/2023/10/14/apple-and-cmu-researchers-unveil-the-never-ending-ui-learner-revolutionizing-app-accessibility-through-continuous-machine-learning/#respond Sun, 15 Oct 2023 02:36:39 +0000 https://www.marktechpost.com/?p=44548 Machine learning is becoming increasingly integrated across a wide range of fields. Its widespread use extends to all industries, including the world of user interfaces (UIs), where it is crucial for anticipating semantic data. This application not only improves accessibility and simplifies testing but also helps automate UI-related tasks, resulting in more streamlined and effective […]

The post Apple and CMU Researchers Unveil the Never-ending UI Learner: Revolutionizing App Accessibility Through Continuous Machine Learning appeared first on MarkTechPost.

]]>

Machine learning is becoming increasingly integrated across a wide range of fields. Its widespread use extends to all industries, including the world of user interfaces (UIs), where it is crucial for anticipating semantic data. This application not only improves accessibility and simplifies testing but also helps automate UI-related tasks, resulting in more streamlined and effective applications.

Currently, many models mainly rely on datasets of static screenshots that humans have rated. But this approach is expensive and exposes unanticipated inclinations toward mistakes in some activities. Because they cannot interact with the UI element in the live app to confirm their conclusions, human annotators must depend solely on visual clues when evaluating if a UI element is tappable from a snapshot.

Despite the drawbacks of using datasets that only record fixed snapshots of mobile application views, they are expensive to use and maintain. However, due to their abundance of data, these datasets continue to be invaluable for training Deep Neural Networks (DNNs). 

Consequently, Apple researchers have developed the Never-Ending UI Learner AI system in collaboration with Carnegie Mellon University. This system interacts continually with actual mobile applications, allowing it to continuously improve its understanding of UI design patterns and new trends. It autonomously downloads apps from app stores for mobile devices and thoroughly investigates each one to find fresh and difficult training scenarios.

The Never-Ending UI Learner has explored over 5,000 device hours so far, performing more than 500,000 actions across 6,000 apps. Due to this prolonged interaction, three different computer vision models will be trained: one for predicting tappability,  another for predicting draggability, and a third for determining screen similarity.

It performs numerous interactions, such as taps and swipes, on components inside the user interface of each app during this research. The researchers emphasize that it classifies UI elements using designed heuristics, identifying characteristics like whether a button may be touched or an image can be moved. 

With the help of the collected data, models that forecast the tappability and draggability of UI elements and the similarity of seen screens are trained. The end-to-end procedure does not require any more human-labeled examples, even if the process can begin with a model trained on human-labeled data.

The researchers emphasized that this method of actively investigating apps has a benefit. It assists the machine in identifying challenging circumstances that typical human-labeled datasets could overlook. Occasionally, people may not notice everything that can be touched on a screen because the images aren’t always very clear. However, the crawler can tap on items and immediately watch what happens, providing clearer and better information.

The researchers demonstrated how models trained on this data improve over time, with tappability prediction reaching 86% accuracy after five training rounds. 

The researchers highlighted that applications focused on accessibility repairs might benefit from more frequent updates to catch subtle changes. On the flip side, longer intervals allowing the accumulation of more significant UI changes could be preferable for tasks like summarizing or mining design patterns. Figuring out the best schedules for retraining and updates will require further research.

This work emphasizes the possibility of never-ending learning, enabling systems to adapt and advance as they take in more data continuously. While the current system focuses on modeling simple semantics like tappability, Apple hopes to apply similar principles to learn more sophisticated representations of mobile UIs and interaction patterns.


Check out the PaperAll Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post Apple and CMU Researchers Unveil the Never-ending UI Learner: Revolutionizing App Accessibility Through Continuous Machine Learning appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/14/apple-and-cmu-researchers-unveil-the-never-ending-ui-learner-revolutionizing-app-accessibility-through-continuous-machine-learning/feed/ 0 44548
Google Researchers Introduce An Open-Source Library in JAX for Deep Learning on Spherical Surfaces https://www.marktechpost.com/2023/10/10/google-researchers-introduce-an-open-source-library-in-jax-for-deep-learning-on-spherical-surfaces/ https://www.marktechpost.com/2023/10/10/google-researchers-introduce-an-open-source-library-in-jax-for-deep-learning-on-spherical-surfaces/#respond Wed, 11 Oct 2023 04:30:00 +0000 https://www.marktechpost.com/?p=44256 Deep learning, a machine learning subset, automatically learns complex representations from the input. Its applications are used in many fields, such as image and speech recognition for language processing, object detection, and medical imaging diagnostics; finance for algorithmic trading and fraud detection; autonomous vehicles using convolutional neural networks for real-time decision-making; and recommendation systems for […]

The post Google Researchers Introduce An Open-Source Library in JAX for Deep Learning on Spherical Surfaces appeared first on MarkTechPost.

]]>

Deep learning, a machine learning subset, automatically learns complex representations from the input. Its applications are used in many fields, such as image and speech recognition for language processing, object detection, and medical imaging diagnostics; finance for algorithmic trading and fraud detection; autonomous vehicles using convolutional neural networks for real-time decision-making; and recommendation systems for personalized content. 

Convolutional neural networks (CNNs) and vision transformers (ViT), two examples of deep learning models for computer vision, analyze signals by assuming planar (flat) regions. Digital photographs, for example, are presented as a grid of pixels on a flat surface. Nonetheless, this data type represents only a fraction of the diverse data encountered in scientific applications.

However, a few things could be improved by processing spherical signals using a planar approach. First, there is a sampling issue, meaning it is impossible to define uniform grids on the sphere—necessary for planar CNNs and ViTs—without significant distortion. Second, rotations frequently confuse signals and local patterns on the sphere.  To ensure that the model learns the features accurately, we need equivariance to 3D rotations. As a result, the model parameters are used more effectively, and training with less data is possible.

Intuitively, both molecular property prediction and climate forecasting problems should benefit from spherical CNNs. The intrinsic properties of molecules are invariant to rotations of the 3D structure (atom positions), so rotation equivariant representations would provide a natural way to encode this symmetry.

Consequently, the researchers have formulated an open-source library in JAX for deep learning on spherical surfaces. It outperforms state-of-the-art results on benchmarks for molecular property prediction and weather forecasting, typically handled by transformers and graph neural networks.

The researchers highlighted that these can solve both the problems of sampling and of robustness to rotation. It does by leveraging spherical convolution and cross-correlation operations. Spherical CNNs offer promising applications in two critical domains: medical research and climate analysis, holding the potential to catalyze transformative advancements for society.

Spherical CNNs present a theoretical advantage in addressing challenges related to predicting chemical properties and understanding climate states. Leveraging rotation-equivariant representations becomes particularly logical in capturing the inherent symmetries of molecular structures, where the properties remain invariant to 3D rotations (atom locations).

Since atmospheric data is naturally displayed on a sphere, spherical CNNs are well suited for this task. They can also effectively manage repeated patterns in such data at various places and orientations.

The researchers said that their models exceed or match neural weather models based on traditional CNNs on a number of weather forecasting benchmarks. The model forecasts the values of several atmospheric variables six hours in advance, and the results from a test environment are shown below. Then, the model is further evaluated up to five days in advance during training and makes predictions up to three days in advance.

Additionally, the models exhibit exceptional performance across various weather forecasting scenarios, demonstrating the effectiveness of spherical CNNs as neural weather models in a ground-breaking accomplishment. This study outlines the best strategies for scaling spherical CNNs and provides real data to support their applicability in these particular applications.


Check out the Paper, Github, and Google BlogAll Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post Google Researchers Introduce An Open-Source Library in JAX for Deep Learning on Spherical Surfaces appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/10/google-researchers-introduce-an-open-source-library-in-jax-for-deep-learning-on-spherical-surfaces/feed/ 0 44256
Oligo Sounds an Alarm: Critical Security Flaw in TorchServe Could Let Hackers Hijack Servers and Inject Malicious AI Models https://www.marktechpost.com/2023/10/10/oligo-sounds-an-alarm-critical-security-flaw-in-torchserve-could-let-hackers-hijack-servers-and-inject-malicious-ai-models/ https://www.marktechpost.com/2023/10/10/oligo-sounds-an-alarm-critical-security-flaw-in-torchserve-could-let-hackers-hijack-servers-and-inject-malicious-ai-models/#respond Tue, 10 Oct 2023 17:12:51 +0000 https://www.marktechpost.com/?p=44216 Large language models and AI models are growing in popularity every day. Even preventing accidents, detecting cancer, and maintaining public safety, we must rely on these AI apps to provide the finest information. Armed forces and weapons are also utilizing AI in international confrontations. Machine learning (ML) research has been largely driven by PyTorch, which […]

The post Oligo Sounds an Alarm: Critical Security Flaw in TorchServe Could Let Hackers Hijack Servers and Inject Malicious AI Models appeared first on MarkTechPost.

]]>

Large language models and AI models are growing in popularity every day. Even preventing accidents, detecting cancer, and maintaining public safety, we must rely on these AI apps to provide the finest information. Armed forces and weapons are also utilizing AI in international confrontations.

Machine learning (ML) research has been largely driven by PyTorch, which stands out as a leading AI platform. Although PyTorch is widely used in studies—over 90% of ML research publications use it—its prominence makes it a tempting target for potential attackers looking to infiltrate AI-based systems. Notably, PyTorch has a wide range of customers, including some of the biggest businesses in the world, including Walmart, Amazon, OpenAI, Tesla, Azure, Google Cloud, Intel, and others.

However, Oligo Security mistakenly discovered that TorchServe’s default configuration could be compromised. Oligo found a brand-new critical SSRF weakness in the administration interface that allowed configuration uploads from any domain and allowed for remote code execution (RCE). An attacker can run code and take control of the target server by taking advantage of ShellTorch. 

They noticed that TorchServe is vulnerable to unsafe deserialization of a malicious model, which could allow remote code execution. This combination of vulnerabilities could lead to Remote Code Execution (RCE) and a complete takeover, especially given the substantial number of TorchServe applications, with tens of thousands of instances being exposed to these risks. They observed that many openly available, unprotected instances are vulnerable to hacking, the introduction of malicious AI models, and even a full server takeover. They emphasized that it may affect millions of people. The world’s servers can be compromised due to these flaws. Therefore, some of the biggest businesses in the world may be immediately in danger.

Consequently, the researchers developed a security product to detect threats within a runtime environment. Unlike other tools that may miss certain causes of undesirable or unsafe application behavior, Oligo investigates the dynamic environment where libraries are utilized, identifying issues that might be overlooked. In contrast to static analysis solutions, it also can spot anomalies in any code during runtime. This includes code developed with open-source libraries, proprietary third-party software, or custom code. Oligo also identifies potential risk sources, such as insecure configuration settings. So, it is possible to see, change, steal, and delete AI models and sensitive data going into and out of the target TorchServe server using the high privileges made available by these vulnerabilities. 

The researchers emphasized that an additional advantage of Oligo is its ability to offer low-disruption solutions. It doesn’t necessarily require comprehensive patching or version changes when addressing vulnerabilities and security issues, providing a more streamlined approach to enhancing system security.


Check out the Reference ArticleAll Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post Oligo Sounds an Alarm: Critical Security Flaw in TorchServe Could Let Hackers Hijack Servers and Inject Malicious AI Models appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/10/oligo-sounds-an-alarm-critical-security-flaw-in-torchserve-could-let-hackers-hijack-servers-and-inject-malicious-ai-models/feed/ 0 44216
Stanford Researchers Propose MAPTree: A Bayesian Approach to Decision Tree Induction with Enhanced Robustness and Performance https://www.marktechpost.com/2023/10/08/stanford-researchers-propose-maptree-a-bayesian-approach-to-decision-tree-induction-with-enhanced-robustness-and-performance/ https://www.marktechpost.com/2023/10/08/stanford-researchers-propose-maptree-a-bayesian-approach-to-decision-tree-induction-with-enhanced-robustness-and-performance/#respond Mon, 09 Oct 2023 04:40:42 +0000 https://www.marktechpost.com/?p=44130 Decision trees are a popular machine learning algorithm that can be used for both classification and regression tasks. They operate by recursively dividing the dataset into subsets according to the most important property at each node. A tree structure illustrates the decision-making process, with each internal node designating a choice based on an attribute, each […]

The post Stanford Researchers Propose MAPTree: A Bayesian Approach to Decision Tree Induction with Enhanced Robustness and Performance appeared first on MarkTechPost.

]]>

Decision trees are a popular machine learning algorithm that can be used for both classification and regression tasks. They operate by recursively dividing the dataset into subsets according to the most important property at each node. A tree structure illustrates the decision-making process, with each internal node designating a choice based on an attribute, each branch standing for the choice’s result, and each leaf node for the result. They are praised for their efficiency, adaptability, and interpretability. 

In a work titled “MAPTree: Surpassing ‘Optimal’ Decision Trees using Bayesian Decision Trees,” a team from Stanford University formulated the MAPTree algorithm. This method determines the maximum a posteriori tree by expertly assessing the posterior distribution of Bayesian Classification and Regression Trees (BCART) created for a specific dataset. The study shows that MAPTree can successfully enhance decision tree models beyond what was previously believed to be optimum.

Bayesian Classification and Regression Trees (BCART) have become an advanced approach, introducing a posterior distribution over tree structures based on available data. This approach, in practice, tends to outshine conventional greedy methods by producing superior tree structures. However, it suffers from the drawback of having exponentially long mixing times and often getting trapped in local minima.

The researchers developed a formal connection between AND/OR search issues and the maximum a posteriori inference of Bayesian Classification and Regression Trees (BCART), illuminating the problem’s fundamental structure. The researchers emphasized that the creation of individual decision trees is the main emphasis of this study. It contests the idea of optimal decision trees, which casts the induction of decision trees as a global optimization problem aimed at maximizing an overall objective function.

As a more sophisticated method, Bayesian Classification and Regression Trees (BCART) provide a posterior distribution across tree architectures based on available data. This method produces superior tree architectures compared to traditional greedy methods. 

The researchers also emphasized that MAPTree offers practitioners faster outcomes by outperforming earlier sampling-based strategies regarding computational efficiency. The trees found by MAPTree performed better than the most advanced algorithms currently available or performed similarly while leaving a lesser environmental footprint. 

They used a collection of 16 datasets from the CP4IM dataset to evaluate the generalization accuracy, log-likelihood, and tree size of models created by MAPTree and the baseline techniques. They found that MAPTree either outperforms the baselines in test accuracy or log-likelihood, or produces noticeably slimmer decision trees in situations of similar performance.

In conclusion, MAPTree offers a quicker, more effective, and more effective alternative to current methodologies, representing a significant advancement in decision tree modeling. Its potential influence on data analysis and machine learning cannot be emphasized, offering professionals a potent tool for building decision trees that excel in performance and efficiency.


Check out the Paper and GithubAll Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post Stanford Researchers Propose MAPTree: A Bayesian Approach to Decision Tree Induction with Enhanced Robustness and Performance appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/08/stanford-researchers-propose-maptree-a-bayesian-approach-to-decision-tree-induction-with-enhanced-robustness-and-performance/feed/ 0 44130
This AI Paper Reveals a New Approach to Understand Deep Learning Models: Unpacking the ‘Where’ and ‘What’ with Concept Relevance Propagation (CRP) https://www.marktechpost.com/2023/10/08/this-ai-paper-reveals-a-new-approach-to-understand-deep-learning-models-unpacking-the-where-and-what-with-concept-relevance-propagation-crp/ https://www.marktechpost.com/2023/10/08/this-ai-paper-reveals-a-new-approach-to-understand-deep-learning-models-unpacking-the-where-and-what-with-concept-relevance-propagation-crp/#respond Sun, 08 Oct 2023 07:49:50 +0000 https://www.marktechpost.com/?p=44072 The field of Machine Learning and Artificial Intelligence has become very important. We have new advancements that have been there with each day. The area is impacting all spheres. By utilizing finely developed neural network architectures, we have models that are distinguished by extraordinary accuracy within their respective sectors. Despite their accurate performance, we must […]

The post This AI Paper Reveals a New Approach to Understand Deep Learning Models: Unpacking the ‘Where’ and ‘What’ with Concept Relevance Propagation (CRP) appeared first on MarkTechPost.

]]>

The field of Machine Learning and Artificial Intelligence has become very important. We have new advancements that have been there with each day. The area is impacting all spheres. By utilizing finely developed neural network architectures, we have models that are distinguished by extraordinary accuracy within their respective sectors.

Despite their accurate performance, we must still fully understand how these neural networks function. We must know the mechanisms governing attribute selection and prediction inside these models to observe and interpret results.

The intricate and nonlinear nature of deep neural networks (DNNs) often leads to conclusions that may exhibit bias towards undesired or undesirable traits. The inherent opacity of their reasoning poses a challenge, making it challenging to apply machine learning models across various relevant application domains. It isn’t easy to understand how an AI system makes its decisions.

Consequently, Prof. Thomas Wiegand (Fraunhofer HHI, BIFOLD), Prof. Wojciech Samek (Fraunhofer HHI, BIFOLD), and Dr. Sebastian Lapuschkin (Fraunhofer HHI) introduced the concept of relevance propagation (CRP) in their paper. This innovative method offers a pathway from attribution maps to human-understandable explanations, allowing for the elucidation of individual AI decisions through concepts understandable to humans.

They highlight CRP as an advanced explanatory method for deep neural networks to complement and enrich existing explanatory models. By integrating local and global perspectives, CRP addresses the ‘where’ and ‘what’ questions about individual predictions. The AI ideas CRP uses, their spatial representation in the input, and the individual neural network segments responsible for their consideration are all revealed by CRP, in addition to the relevant input variables impacting the choice.

As a result, CRP describes decisions made by AI in terms that people can comprehend. 

The researchers emphasize that this approach of explainability examines an AI’s full prediction process from input to output. The research group has already created techniques for using heat maps to demonstrate how AI algorithms make judgments.

Dr. Sebastian Lapuschkin, head of the research group Explainable Artificial Intelligence at Fraunhofer HHI, explains the new technique in more detail. He said that CRP transfers the explanation from the input space, where the image with all its pixels is located, to the semantically enriched concept space formed by higher neural network layers. 

The researchers further said that the next phase of AI explainability, known as CRP, opens up a world of new opportunities for researching, evaluating, and enhancing the performance of AI models.

Insights into the representation and composition of ideas within the model and a quantitative evaluation of their influence on predictions can be acquired by exploring model designs and application domains using CRP-based studies. These investigations leverage the power of CRP to delve into the intricate layers of the model, unraveling the conceptual landscape and assessing the quantitative impact of various ideas on predictive outcomes. 


Check out the PaperAll Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 31k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.

If you like our work, you will love our newsletter..

We are also on WhatsApp. Join our AI Channel on Whatsapp..

The post This AI Paper Reveals a New Approach to Understand Deep Learning Models: Unpacking the ‘Where’ and ‘What’ with Concept Relevance Propagation (CRP) appeared first on MarkTechPost.

]]>
https://www.marktechpost.com/2023/10/08/this-ai-paper-reveals-a-new-approach-to-understand-deep-learning-models-unpacking-the-where-and-what-with-concept-relevance-propagation-crp/feed/ 0 44072