Menu

Blog

Archive for the ‘information science’ category: Page 151

Jan 6, 2022

Parkinson’s Drug Discovery Collaboration Between Astrogen, Iktos to Leverage AI Platform

Posted by in categories: biotech/medical, chemistry, information science, robotics/AI

Artificial intelligence drug design company Iktos, and South Korean clinical research biotech Astrogen announced today a collaboration with the goal of discovering small-molecule pre-clinical drug candidates for a specific, undisclosed, marker of Parkinson’s disease (PD).

Under the terms of the agreement, whose value was not disclosed, Iktos will apply its generative learning algorithms which seek to identify new molecular structures with the potential address the target in PD. Astrogen, which has a focus of the development of therapeutics for “intractable neurological diseases,” will provide in-vitro and in-vivo screening of lead compounds and pre-clinical compounds. While both companies will contribute to the identification of new small-molecule candidates, Astrgoen will lead the drug development process from the pre-clinical stages.

“Our objective is to expedite drug discovery and achieve time and cost efficiencies for our global collaborators by using Iktos’s proprietary AI platform and know-how,” noted Yann Gaston-Mathé, president and CEO of Paris-based Iktos in a press release. “We are confident that together we will be able to identify promising novel chemical matter for the treatment of intractable neurological diseases. Our strategy has always been to tackle challenging problems alongside our collaborators where we can demonstrate value generation for new and on-going drug discovery projects.”

Jan 3, 2022

Artificial Intelligence of the Future Could Reveal the Incomprehensible

Posted by in categories: biotech/medical, information science, robotics/AI

The research study of Spanish clinical neuropsychologist Gabriel G. De la Torre, Does artificial intelligence dream of non-terrestrial techno-signatures?, suggests that one of the “potential applications of artificial intelligence is not only to assist in big data analysis but to help to discern possible artificiality or oddities in patterns of either radio signals, megastructures or techno-signatures in general.”

“Our form of life and intelligence,” observed Silvano P. Colombano at NASA’s Ames Research Center who was not involved in the study, “may just be a tiny first step in a continuing evolution that may well produce forms of intelligence that are far superior to ours and no longer based on carbon ” machinery.”

Jan 1, 2022

Preparation is key to AI success in 2022

Posted by in categories: business, information science, robotics/AI

Artificial intelligence is unlike previous technology innovations in one crucial way: it’s not simply another platform to be deployed, but a fundamental shift in the way data is used. As such, it requires a substantial rethinking as to the way the enterprise collects, processes, and ultimately deploys data to achieve business and operational objectives.

So while it may be tempting to push AI into legacy environments as quickly as possible, a wiser course of action would be to adopt a more careful, thoughtful approach. One thing to keep in mind is that AI is only as good as the data it can access, so shoring up both infrastructure and data management and preparation processes will play a substantial role in the success or failure of future AI-driven initiatives.

According to Open Data Science, the need to foster vast amounts of high-quality data is paramount for AI to deliver successful outcomes. In order to deliver valuable insights and enable intelligent algorithms to continuously learn, AI must connect with the right data from the start. Not only should organizations develop sources of high-quality data before investing in AI, but they should also reorient their entire cultures so that everyone from data scientists to line-of-business knowledge workers understand the data needs of AI and how results can be influenced by the type and quality of data being fed into the system.

Jan 1, 2022

Predicting the Difficulty of Texts Using Machine Learning and Getting a Visual Representation of Words

Posted by in categories: information science, robotics/AI

We see that text data is ubiquitous in nature. There is a lot of text present in different forms such as posts, books, articles, and blogs. What is more interesting is the fact that there is a subset of Artificial Intelligence called Natural Language Processing (NLP) that would convert text into a form that could be used for machine learning. I know that sounds a lot but getting to know the details and the proper implementation of machine learning algorithms could ensure that one learns the important tools in the process.

Since the r e are newer and better libraries being created to be used for machine learning purposes, it would make sense to learn some of the state-of-the-art tools that could be used for predictions. I’ve recently come across a challenge on Kaggle about predicting the difficulty of the text.

The output variable, the difficulty of the text, is converted into a form that is continuous in nature. This makes the target variable continuous. Therefore, various regression techniques must be used for predicting the difficulty of the text. Since the text is ubiquitous in nature, applying the right processing mechanisms and predictions would be really valuable, especially for companies that receive feedback and reviews in the form of text.

Dec 30, 2021

Is social media killing intellectual humility?

Posted by in categories: education, information science, internet

An echo chamber is an infinity of mirrors. Photo: Robert Brook via Getty Images

“One way the internet distorts our picture of ourselves is by feeding the human tendency to overestimate our knowledge of how the world works,” writes philosophy professor Michael Patrick Lynch, author of the book The Internet of Us: Knowing More and Understanding Less in the Age of Big Data, in The Chronicle of Higher Education. “The Internet of Us becomes one big reinforcement mechanism, getting us all the information we are already biased to believe, and encouraging us to regard those in other bubbles as misinformed miscreants. We know it all—the internet tells us so.”

Dec 30, 2021

Social Network for Programmers and Developers

Posted by in category: information science

Social network for developers to discuss topics about bugs and issues, write and share knowledge and connect with millions of developers worldwide.

Dec 29, 2021

Simple, accurate, and efficient: Improving the way computers recognize hand gestures

Posted by in categories: information science, mobile phones, robotics/AI, wearables

In the 2002 science fiction blockbuster film “Minority Report,” Tom Cruise’s character John Anderton uses his hands, sheathed in special gloves, to interface with his wall-sized transparent computer screen. The computer recognizes his gestures to enlarge, zoom in, and swipe away. Although this futuristic vision for computer-human interaction is now 20 years old, today’s humans still interface with computers by using a mouse, keyboard, remote control, or small touch screen. However, much effort has been devoted by researchers to unlock more natural forms of communication without requiring contact between the user and the device. Voice commands are a prominent example that have found their way into modern smartphones and virtual assistants, letting us interact and control devices through speech.

Hand gestures constitute another important mode of human communication that could be adopted for human-computer interactions. Recent progress in camera systems, image analysis and machine learning have made optical-based gesture recognition a more attractive option in most contexts than approaches relying on wearable sensors or data gloves, as used by Anderton in “Minority Report.” However, current methods are hindered by a variety of limitations, including high computational complexity, low speed, poor accuracy, or a low number of recognizable gestures. To tackle these issues, a team led by Zhiyi Yu of Sun Yat-sen University, China, recently developed a new hand gesture recognition algorithm that strikes a good balance between complexity, accuracy, and applicability.

Dec 27, 2021

New AI improves itself through Darwinian-style evolution

Posted by in categories: information science, mathematics, robotics/AI

AutoML-Zero is unique because it uses simple mathematical concepts to generate algorithms “from scratch,” as the paper states. Then, it selects the best ones, and mutates them through a process that’s similar to Darwinian evolution.

AutoML-Zero first randomly generates 100 candidate algorithms, each of which then performs a task, like recognizing an image. The performance of these algorithms is compared to hand-designed algorithms.-Zero then selects the top-performing algorithm to be the “parent.”

“This parent is then copied and mutated to produce a child algorithm that is added to the population, while the oldest algorithm in the population is removed,” the paper states.

Dec 24, 2021

Heart Rate Detection using Eulerian Magnification + YOLOR

Posted by in categories: drones, information science, robotics/AI

Real Time Heart Rate Detection using Eulerian Magnification + YOLOR is used for head detection which feeds into a Eulerian Magnification algorithm developed by Rohin Tangirala. Courtesy of Dragos Stan for assistance in this demo and code.

⭐️Code+Dataset — https://lnkd.in/deRj6SPf.

Continue reading “Heart Rate Detection using Eulerian Magnification + YOLOR” »

Dec 24, 2021

The quantum mechanics of time travel through post-selected teleportation

Posted by in categories: information science, quantum physics, time travel

This paper discusses the quantum mechanics of closed timelike curves (CTC) and of other potential methods for time travel. We analyze a specific proposal for such quantum time travel, the quantum description of CTCs based on post-selected teleportation (P-CTCs). We compare the theory of P-CTCs to previously proposed quantum theories of time travel: the theory is physically inequivalent to Deutsch’s theory of CTCs, but it is consistent with path-integral approaches (which are the best suited for analyzing quantum field theory in curved spacetime). We derive the dynamical equations that a chronology-respecting system interacting with a CTC will experience. We discuss the possibility of time travel in the absence of general relativistic closed timelike curves, and investigate the implications of P-CTCs for enhancing the power of computation.