Thrust 5: Streetscape Applications
Design of an application services architecture that will enable the exploration of cross-layer or thrust optimizations and support novel streetscape applications.
Project 1: Human-Computer Interaction Design for Smart Streetscapes
Streetscape applications are fundamentally unique in their transparent integration within public spaces. They are not equipped with traditional displays, input mechanisms, or facilities to opt-in or opt-out. This creates an exciting new opportunity to reimagine human-computer interaction for this new class of system. CS3 is developing, piloting, and evaluating new interaction modalities to enable pedestrians to effectively engage with emerging streetscape applications.
This work explores optimal timing strategies for agent-human interaction in streetscape environments, focusing on seamless, context-aware interventions that enhance pedestrian engagement without disrupting natural movement. By leveraging insights from sensor data and real-time analytics, the system aims to identify precise moments for interaction, such as providing information or safety alerts, based on pedestrian speed, location, and environmental conditions. This approach helps create a more intuitive interface for public spaces, enhancing user experience and safety through minimally intrusive but timely interactions.
This work explores using large language models (LLMs) to analyze extensive urban datasets, demonstrated through a case study on NYC's 311 service requests. By combining data-driven insights with community narratives, the system aims to improve civic engagement and policymaking. Initial results with ChatGPT-4 highlighted challenges with recognizing long-term patterns due to context limitations. The paper suggests solutions, including narrative structure frameworks, spatiotemporal indexing, and fact-checking agents, to generate reliable, context-rich narratives. These enhancements aim to enable LLMs to provide meaningful urban insights, integrating personal experiences and structured data for better community-focused decision-making.
Project 2: Streetscape Application Stack and Runtime Design
Desktop and mobile applications benefit from robust execution substrates provided by underlying runtime platforms and application libraries and services. This has dramatically accelerated application development, broadened the community of developers, and provided performance and security benefits. CS3 is developing the equivalent platforms, libraries, and services for future streetscape applications. A key objective is to establish a new developer community anchored around these enabling (open-source) technologies.
Many CS3 applications will rely upon datasets that live beyond the administrative boundaries of the CS3 team. As one example, in the case of the Smart Intersection application, the application requires access to traffic data beyond the intersection instrumented by CS3. The team is developing standard toolsets to ingest, quality check, and integrate datasets like this (and beyond). Data from NYC department of transportation video streams, and from the Florida testbed is actively being integrated.
Robust decisions in streetscapes and related applications require high fidelity estimation of states and processes in the environment. This activity proposes communication-efficient distributed algorithms to combine sets of cooperative heterogeneous sensors under bandwidth limitation, with provable guarantees. The overarching goals are to manage limited communication bandwidth without suffering performance loss.
Many CS3 applications, including those designed by partners external to the team, depend only upon anonymized, synoptic information about the streetscape. Our Mobility Intelligence application, for example, is concerned only about the objects (e.g., pedestrians, scooters) and activities (e.g., walking, turning left) identified within the streetscape – not the specific characteristics of the participants (e.g., visual information). CS3 is developing a privacy-preserving scene and activity recognition pipeline that simplifies the development of these applications, including those built by external partners. This supports privacy preservation while broadening access for streetscape application developers.
This research is to develop a digital twin platform to support smart streetscape application testing. The digital twin platform is built based on high-resolution LiDAR or photogrammetry scan of the infrastructure and road environment. Trajectories of vehicles, pedestrians, and other road users generated from high-resolution LiDAR and computer vision sensors are used to reconstruct the real-world traffic conditions in the high-resolution infrastructure background. Key research problems lie in addressing the spatial-temporal synchronization and matching of multi-sensor data, predict and infer trajectory data due to line-of-sight coverage loss, and the next-generation traffic simulation models that can simulate high-fidelity traffic behavior in responses to the comprehensive 3D environment in the new digital twin models.
Project 3: Smart Streetscape Operating System or Hypervisor Design
The shift from dedicated computer and network systems (e.g., desktops) to shared-use computer and network systems (e.g., cloud) triggered important changes in the underlying operating systems and ushered in an era of system virtualization, enabling many logical systems to transparently share a smaller number of physical systems. The hypervisors that support this virtualization are a cornerstone of resource utilization efficiency and application scalability. CS3 is exploring new mechanisms –equivalent to components of modern operating systems and hypervisors— to extend these same benefits to future multi-tenant streetscape systems.
Implement a synchronization algorithm that aligns diverse sensors, including cameras and IMUs, to ensure consistent data integration for pedestrian safety applications.
Create a lightweight neural network, CoordinateTransformNet, to convert coordinates from different street-level cameras to a unified top-down view. This facilitates real-time integration of multiple camera feeds for comprehensive pedestrian and vehicle detection at intersections.
Utilize deep learning models such as YOLOv8 for object detection and OC-SORT for tracking. Fine-tune models for urban scenarios to handle various object sizes and complex dynamics, ensuring robust performance even during occlusions or temporary detection failures.