Menu

Blog

Archive for the ‘information science’ category: Page 146

Jan 1, 2022

Preparation is key to AI success in 2022

Posted by in categories: business, information science, robotics/AI

Artificial intelligence is unlike previous technology innovations in one crucial way: it’s not simply another platform to be deployed, but a fundamental shift in the way data is used. As such, it requires a substantial rethinking as to the way the enterprise collects, processes, and ultimately deploys data to achieve business and operational objectives.

So while it may be tempting to push AI into legacy environments as quickly as possible, a wiser course of action would be to adopt a more careful, thoughtful approach. One thing to keep in mind is that AI is only as good as the data it can access, so shoring up both infrastructure and data management and preparation processes will play a substantial role in the success or failure of future AI-driven initiatives.

According to Open Data Science, the need to foster vast amounts of high-quality data is paramount for AI to deliver successful outcomes. In order to deliver valuable insights and enable intelligent algorithms to continuously learn, AI must connect with the right data from the start. Not only should organizations develop sources of high-quality data before investing in AI, but they should also reorient their entire cultures so that everyone from data scientists to line-of-business knowledge workers understand the data needs of AI and how results can be influenced by the type and quality of data being fed into the system.

Jan 1, 2022

Predicting the Difficulty of Texts Using Machine Learning and Getting a Visual Representation of Words

Posted by in categories: information science, robotics/AI

We see that text data is ubiquitous in nature. There is a lot of text present in different forms such as posts, books, articles, and blogs. What is more interesting is the fact that there is a subset of Artificial Intelligence called Natural Language Processing (NLP) that would convert text into a form that could be used for machine learning. I know that sounds a lot but getting to know the details and the proper implementation of machine learning algorithms could ensure that one learns the important tools in the process.

Since the r e are newer and better libraries being created to be used for machine learning purposes, it would make sense to learn some of the state-of-the-art tools that could be used for predictions. I’ve recently come across a challenge on Kaggle about predicting the difficulty of the text.

The output variable, the difficulty of the text, is converted into a form that is continuous in nature. This makes the target variable continuous. Therefore, various regression techniques must be used for predicting the difficulty of the text. Since the text is ubiquitous in nature, applying the right processing mechanisms and predictions would be really valuable, especially for companies that receive feedback and reviews in the form of text.

Dec 30, 2021

Is social media killing intellectual humility?

Posted by in categories: education, information science, internet

An echo chamber is an infinity of mirrors. Photo: Robert Brook via Getty Images

“One way the internet distorts our picture of ourselves is by feeding the human tendency to overestimate our knowledge of how the world works,” writes philosophy professor Michael Patrick Lynch, author of the book The Internet of Us: Knowing More and Understanding Less in the Age of Big Data, in The Chronicle of Higher Education. “The Internet of Us becomes one big reinforcement mechanism, getting us all the information we are already biased to believe, and encouraging us to regard those in other bubbles as misinformed miscreants. We know it all—the internet tells us so.”

Dec 30, 2021

Social Network for Programmers and Developers

Posted by in category: information science

Social network for developers to discuss topics about bugs and issues, write and share knowledge and connect with millions of developers worldwide.

Dec 29, 2021

Simple, accurate, and efficient: Improving the way computers recognize hand gestures

Posted by in categories: information science, mobile phones, robotics/AI, wearables

In the 2002 science fiction blockbuster film “Minority Report,” Tom Cruise’s character John Anderton uses his hands, sheathed in special gloves, to interface with his wall-sized transparent computer screen. The computer recognizes his gestures to enlarge, zoom in, and swipe away. Although this futuristic vision for computer-human interaction is now 20 years old, today’s humans still interface with computers by using a mouse, keyboard, remote control, or small touch screen. However, much effort has been devoted by researchers to unlock more natural forms of communication without requiring contact between the user and the device. Voice commands are a prominent example that have found their way into modern smartphones and virtual assistants, letting us interact and control devices through speech.

Hand gestures constitute another important mode of human communication that could be adopted for human-computer interactions. Recent progress in camera systems, image analysis and machine learning have made optical-based gesture recognition a more attractive option in most contexts than approaches relying on wearable sensors or data gloves, as used by Anderton in “Minority Report.” However, current methods are hindered by a variety of limitations, including high computational complexity, low speed, poor accuracy, or a low number of recognizable gestures. To tackle these issues, a team led by Zhiyi Yu of Sun Yat-sen University, China, recently developed a new hand gesture recognition algorithm that strikes a good balance between complexity, accuracy, and applicability.

Dec 27, 2021

New AI improves itself through Darwinian-style evolution

Posted by in categories: information science, mathematics, robotics/AI

AutoML-Zero is unique because it uses simple mathematical concepts to generate algorithms “from scratch,” as the paper states. Then, it selects the best ones, and mutates them through a process that’s similar to Darwinian evolution.

AutoML-Zero first randomly generates 100 candidate algorithms, each of which then performs a task, like recognizing an image. The performance of these algorithms is compared to hand-designed algorithms.-Zero then selects the top-performing algorithm to be the “parent.”

“This parent is then copied and mutated to produce a child algorithm that is added to the population, while the oldest algorithm in the population is removed,” the paper states.

Dec 24, 2021

Heart Rate Detection using Eulerian Magnification + YOLOR

Posted by in categories: drones, information science, robotics/AI

Real Time Heart Rate Detection using Eulerian Magnification + YOLOR is used for head detection which feeds into a Eulerian Magnification algorithm developed by Rohin Tangirala. Courtesy of Dragos Stan for assistance in this demo and code.

⭐️Code+Dataset — https://lnkd.in/deRj6SPf.

Continue reading “Heart Rate Detection using Eulerian Magnification + YOLOR” »

Dec 24, 2021

The quantum mechanics of time travel through post-selected teleportation

Posted by in categories: information science, quantum physics, time travel

This paper discusses the quantum mechanics of closed timelike curves (CTC) and of other potential methods for time travel. We analyze a specific proposal for such quantum time travel, the quantum description of CTCs based on post-selected teleportation (P-CTCs). We compare the theory of P-CTCs to previously proposed quantum theories of time travel: the theory is physically inequivalent to Deutsch’s theory of CTCs, but it is consistent with path-integral approaches (which are the best suited for analyzing quantum field theory in curved spacetime). We derive the dynamical equations that a chronology-respecting system interacting with a CTC will experience. We discuss the possibility of time travel in the absence of general relativistic closed timelike curves, and investigate the implications of P-CTCs for enhancing the power of computation.

Dec 24, 2021

LightOn Photonic Co-processor Integrated Into European AI Supercomputer

Posted by in categories: information science, robotics/AI, supercomputing

PARIS, Dec. 23, 2021 – LightOn announces the integration of one of its photonic co-processors in the Jean Zay supercomputer, one of the Top500 most powerful computers in the world. Under a pilot program with GENCI and IDRIS, the insertion of a cutting-edge analog photonic accelerator into High Performance Computers (HPC) represents a technological breakthrough and a world-premiere. The LightOn photonic co-processor will be available to selected users of the Jean Zay research community over the next few months.

LightOn’s Optical Processing Unit (OPU) uses photonics to speed up randomized algorithms at a very large scale while working in tandem with standard silicon CPU and NVIDIA latest A100 GPU technology. The technology aims to reduce the overall computing time and power consumption in an area that is deemed “essential to the future of computational science and AI for Science” according to a 2021 U.S. Department of Energy report on “Randomized Algorithms for Scientific Computing.”

INRIA (France’s Institute for Research in Computer Science and Automation) researcher Dr. Antoine Liutkus provided additional context to the integration of LightOn’s coprocessor in the Jean Zay supercomputer: “Our research is focused today on the question of large-scale learning. Integrating an OPU in one of the most powerful nodes of Jean Zay will give us the keys to carry out this research, and will allow us to go beyond a simple ” proof of concept.”

Dec 24, 2021

Azure AI milestone: Microsoft KEAR surpasses human performance on CommonsenseQA benchmark

Posted by in categories: food, information science, robotics/AI

KEAR (Knowledgeable External Attention for commonsense Reasoning) —along with recent milestones in computer vision and neural text-to-speech —is part of a larger Azure AI mission to provide relevant, meaningful AI solutions and services that work better for people because they better capture how people learn and work—with improved vision, knowledge understanding, and speech capabilities. At the center of these efforts is XYZ-code, a joint representation of three cognitive attributes: monolingual text (X), audio or visual sensory signals (Y), and multilingual (Z). For more information about these efforts, read the XYZ-code blog post.

Last month, our Azure Cognitive Services team, comprising researchers and engineers with expertise in AI, achieved a groundbreaking milestone by advancing commonsense language understanding. When given a question that requires drawing on prior knowledge and five answer choices, our latest model— KEAR, Knowledgeable External Attention for commonsense Reasoning —performs better than people answering the same question, calculated as the majority vote among five individuals. KEAR reaches an accuracy of 89.4 percent on the CommonsenseQA leaderboard compared with 88.9 percent human accuracy. While the CommonsenseQA benchmark is in English, we follow a similar technique for multilingual commonsense reasoning and topped the X-CSR leaderboard.

Continue reading “Azure AI milestone: Microsoft KEAR surpasses human performance on CommonsenseQA benchmark” »