February 11, 2022 By Open Innovation Community 3 min read

Find out about the hackathon that brought together contestants who learned new technologies around data science and machine learning, and published their projects on the Cloud Pak for Data Gallery.

The IBM Academy of Technology is made up of enthusiastic IBMers who get together to work on activities of their own choice that they think will be useful to themselves and to IBM. The Best of Data Science & Machine Learning (ML) Projects Hackathon, led by Thomas Schäck, Distinguished Engineer for Watson Studio, is an example. In 2021, the hackathon brought together contestants who learned new technologies and put them into practice, extended their network, and located colleagues with suitable expertise – and they had fun too.


Two maps showing results of flood risks from the data science and machine learning hackathon project.

From the competing projects, two were published on the Cloud Pak for Data Gallery:

  • 1. Flood risk :
    Floods are the most frequent type of natural disaster and can cause widespread devastation, resulting in loss of life and damages to personal property and critical public health infrastructure. Flooding occurs in every U.S. state and territory, and is a threat experienced anywhere in the world that receives rain. According to NOAA, in the U.S. floods kill more people each year than tornadoes, hurricanes, or lightning. Understanding flood risk is important so that people don’t ignore the warnings sent out by agencies like the National Weather Service (NWS). And warnings need to be more specific, pinpointing certain areas and exposed locations. This project extrapolates the FAIR Model for flood analysis. FAIR, short for “Factor Analysis of Information Risk” is the only international standard quantitative model for information security and operational risk.  As described in wikipedia, FAIR underlines that risk is an uncertain event and one should not focus on what is possible, but on how probable a given event is. This probabilistic approach is applied to every factor that is analyzed. The risk is the probability of a loss tied to an asset. In FAIR, risk is defined as the “probable frequency and probable magnitude of future loss”. FAIR further decomposes risk by breaking down different factors that make up probable frequency and probable loss. These factors include: Threat Event Frequency, Vulnerability, Threat Capability, Primary Loss Magnitude, Secondary Risk.  The project calculates the Loss Event Frequency based on how vulnerable and susceptible the flood location is. And finally, based on the severity of the Flood alert and Loss Event Frequency, the Final Threat Level is calculated.
  • 2. Site search :
    Site search recommender improves search relevancy by using user behavior data from ibm.com search and de-identified for public consumption. It’s built using open-source deep learning libraries (TensorFlow and Keras) and implements the collaborative filtering algorithm to make meaningful recommendations to users based on their search data terms and historical search behavior. Benefits of this project include allowing data scientists to improve relevancy of corporate site search results, serving as boilerplate to provide out-of-box support for search use case and leverages data and AI to solve real-life search and discovery challenges.

You can try the projects Flood risk  and Site search yourself. Note:

With many thanks to the data science community in the IBM Academy of Technology for their energy, dedication, and determination and to the two teams who created these projects.

Authors:

Thomas Schaeck, schaeck@de.ibm.com

Susan Malaika, malaika@us.ibm.com


The content in this blog post is the opinion of the author. For more on the IBM Academy of Technology, see these posts:

A Path to the Open Organization – Academy of Technology President Julie Schuneman

About the Academy

Was this article helpful?
YesNo

More from Cloud

How fintechs are helping banks accelerate innovation while navigating global regulations

4 min read - Financial institutions are partnering with technology firms—from cloud providers to fintechs—to adopt innovations that help them stay competitive, remain agile and improve the customer experience. However, the biggest hurdle to adopting new technologies is security and regulatory compliance. While third and fourth parties have the potential to introduce risk, they can also be the solution. As enterprises undergo their modernization journeys, fintechs are redefining digital transformation in ways that have never been seen before. This includes using hybrid cloud and…

IBM Cloud expands its VPC operations in Dallas, Texas

3 min read - Everything is bigger in Texas—including the IBM Cloud® Network footprint. Today, IBM Cloud opened its 10th data center in Dallas, Texas, in support of their virtual private cloud (VPC) operations. DAL14, the new addition, is the fourth availability zone in the IBM Cloud area of Dallas, Texas. It complements the existing setup, which includes two network points of presence (PoPs), one federal data center, and one single-zone region (SZR). The facility is designed to help customers use technology such as…

Apache Kafka use cases: Driving innovation across diverse industries

6 min read - Apache Kafka is an open-source, distributed streaming platform that allows developers to build real-time, event-driven applications. With Apache Kafka, developers can build applications that continuously use streaming data records and deliver real-time experiences to users. Whether checking an account balance, streaming Netflix or browsing LinkedIn, today’s users expect near real-time experiences from apps. Apache Kafka’s event-driven architecture was designed to store data and broadcast events in real-time, making it both a message broker and a storage unit that enables real-time…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters