Fundamental to automating decision making in many application domains is the ability to analyze Big Data. However, maximizing the value of analyzing big data is the need for analytics techniques to be able to “see” beyond” just syntactic representations of data. Rather, the meaning of data or data semantics must be represented in a machine-processible manner so that the exact nature of relationships in data can be exploited effectively. In particular, many applications require the assembly of different kinds of heterogeneous data to be analyzed to solve important problems. This is impossible without the ability to automatically reason about the way in which the different data are related. Data semantics is also being considered as a critical component of enabling “interpretability” of the outcomes of machine and deep learning techniques which currently are mostly semantics-oblivious.
Many modern network security incidents originate from the Web. For instance, it is not uncommon for users to stumble upon a website that hosts malicious advertisements, which in turn may redirect to phishing sites or promote the installation of malicious software via social engineering attacks. In corporate networks, such attacks can have devastating consequences. For example, an initial web-driven malware infection may be used as a stepping stone for larger scale network intrusions and costly data breaches. When such high-profile incidents are discovered, often weeks or even months after the initial attack took place, a digital forensics team is typically called in to reconstruct the root causes of the incident, so that better network defenses and security policies can be developed. However, a forensic analyst may not be able to reconstruct the entire chain of events up to the initial web attack that is the true root cause of the network breach. This is because modern browsers lack the ability to produce detailed audit logs, and the information contained in the existing navigation history and browser cache is typically too sparse or short lived to allow for a detailed reconstruction of complex web attacks.
The Web has redefined the way in which malicious activities such as online harassment, disinformation, and radicalization are carried out. To be able to fully understand these phenomena, we need computational tools able to trace malicious activity as it happens and identify influential entities that carry it out. In this talk, I will present our efforts in developing tools to automatically monitor and model malicious online activities such as coordinated aggression and disinformation. I will then discuss possible mitigations against these harmful activities, keeping in mind the potential unintended consequences that might arise from suspending offending users.
Applied Artificial Intelligence research and platform development for Cybersecurity & Environmental Management supporting Department of Defense and Department of Energy problem-sets
I will share AI research work at ARC supporting the Department of Defense and the Department of Energy focused on Cybersecurity and Environmental Management. This will provide an overview of research & platform development at ARC with possible collaboration with SCIS.
HBCUs were founded to educate primarily African-American students in the United States and since 1837 have continued to provide quality education, culture, and social environments for Black students to thrive and create their own identities. These institutions have produced one third of all Black STEM PhD recipients, produced 19% of all STEM Bachelor’s degrees, and provided clear pathways to the middle class. HBCU researchers and alumni describe these institutions as places providing “freedom to explore”, “be yourself and develop yourself”, and “unapologetic Black spaces”. HBCUs are rich with social capital crucial to nurturing and promoting academic success.
New research ideas require an instrument where they can be developed, tested – and shared. To support Computer Science research such an instrument has to provide access to a diversity of hardware configurations, support deployment at scale, and deep reconfigurability so that a wide range of experiments can be supported. It also has to provide mechanisms for sharing so that new results can multiply their value by triggering further innovation. Most importantly — since science does not stand still – it requires constant adaptation to support an ever increasing range of experiments driven by emergent ideas and opportunities.