Video-based person re-identification (re-id) is an important application in practice. Since large variations exist between different pedestrian videos, as well as within each video, it is challenging ...to conduct re-id between the pedestrian videos. In this paper, we propose a simultaneous intra-video and inter-video distance learning (SI 2 DL) approach for the video-based person re-id. Specifically, SI 2 DL simultaneously learns an intra-video distance metric and an inter-video distance metric from the training videos. The intra-video distance metric is used to make each video more compact, and the inter-video one is used to ensure that the distance between truly matching videos is smaller than that between wrong matching videos. Considering that the goal of distance learning is to make truly matching video pairs from different persons be well separated with each other, we also propose a pair separation-based SI 2 DL (P-SI 2 DL). P-SI 2 DL aims to learn a pair of distance metrics, under which any two truly matching video pairs can be well separated. Experiments on four public pedestrian image sequence data sets show that our approaches achieve the state-of-the-art performance.
The first video cassette recorders were promoted in the 1970s as an extension of broadcast television technology--a time-shifting device, a way to tape TV shows. Early advertising for Sony's Betamax ...told potential purchasers "You don't have to miss Kojak because you're watching Columbo." But within a few years, the VCR had been transformed from a machine that recorded television into an extension of the movie theater into the home. This was less a physical transformation than a change in perception, but one that relied on the very tangible construction of a network of social institutions to support this new marketplace for movies. In From Betamax to Blockbuster, Joshua Greenberg explains how the combination of neighborhood video stores and the VCR created a world in which movies became tangible consumer goods. Greenberg charts a trajectory from early "videophile" communities to the rise of the video store--complete with theater marquee lights, movie posters, popcorn, and clerks who offered expert advice on which movies to rent. The result was more than a new industry; by placing movies on cassette in the hands (and control) of consumers, video rental and sale led to a renegotiation of the boundary between medium and message, and ultimately a new relationship between audiences and movies. Eventually, Blockbuster's top-down franchise store model crowded local video stores out of the market, but the recent rise of Netflix, iTunes, and other technologies have reopened old questions about what a movie is and how (and where) it ought to be watched. By focusing on the "spaces in between" manufacturers and consumers, Greenberg's account offers a fresh perspective on consumer technology, illustrating how the initial transformation of movies from experience into commodity began not from the top down or the bottom up, but from the middle of the burgeoning industry out.Joshua M. Greenberg is Director of Digital Strategy and Scholarship at the New York Public Library.
Vidding is a well-established remix practice where fans edit an existing film, music video, TV show, or other performance and set it to music of their choosing. Vids emerged forty years ago as a ...complicated technological feat involving capturing footage from TV with a VCR and syncing with music—and their makers and consumers were almost exclusively women, many of them queer women. The technological challenges of doing this kind of work in the 1970s and 1980s when vidding began gave rise to a rich culture of collective work, as well as conventions of creators who gathered to share new work and new techniques. While the rise of personal digital technology eventually democratized the tools vidders use, the collective aspect of the culture grew even stronger with the advent of YouTube, Vimeo, and other channels for sharing work. Vidding: A History emphasizes vidding as a critical, feminist form of fan practice. Working outward from interviews, VHS liner notes, convention programs, and mailing list archives, Coppa offers a rich history of vidding communities as they evolved from the 1970s through to the present. Built with the classroom in mind, the open-access electronic version of this book includes over one-hundred vids and an appendix that includes additional close readings of vids.
Codename Revolution Jones, Steven E; Thiruvathukal, George K
2012, 20120224, 2012-02-24, 20120101
eBook
The Nintendo Wii, introduced in 2006, helped usher in a moment of retro-reinvention in video game play. This hugely popular console system, codenamed Revolution during development, signaled a turn ...away from fully immersive, time-consuming MMORPGs or forty-hour FPS games and back toward family fun in the living room. Players using the wireless motion-sensitive controller (the Wii Remote, or "Wiimote") play with their whole bodies, waving, swinging, swaying. The mimetic interface shifts attention from what's on the screen to what's happening in physical space. This book describes the Wii's impact in technological, social, and cultural terms, examining the Wii as a system of interrelated hardware and software that was consciously designed to promote social play in physical space. Each chapter of Codename Revolution focuses on a major component of the Wii as a platform: the console itself, designed to be low-powered and nimble; the iconic Wii Remote; Wii Fit Plus , and its controller, the Wii Balance Board; the Wii Channels interface and Nintendo's distribution system; and the Wii as a social platform that not only affords multiplayer options but also encourages social interaction in shared physical space. Finally, the authors connect the Wii's revolution in mimetic interface gaming--which eventually led to the release of Sony's Move and Microsoft's Kinect--to some of the economic and technological conditions that influence the possibility of making something new in this arena of computing and culture.
Video compression is the core technology in mobile (mHealth) and electronic (eHealth) health video streaming applications. With global video traffic projected to reach 82% of all Internet traffic by ...2022, there is a strong need to develop efficient compression algorithms to accommodate expected future growth. For the first time in decades, and especially since ISO/IEC MPEG and ITU-T VCEG expert groups strategically joined forces to develop the highly successful H.264/AVC standard, we have two distinct initiatives competing for the best performing video codec. On the one hand, we have the Alliance for Open Media (AOM) that support a new, royalty free video codec generation, termed AV1, based on VP8 and VP9 efforts. On the other hand, the Joint Video Exploration Team (JVET) has been developing the Versatile Video Codec (VVC) as the successor of the High Efficiency Video Coding (HEVC) standard. At the same time, the breadth of applications utilizing video codecs, involving significant content variability and moving across the video resolution ladder, to satisfy different constraints, have resulted in mixed literature results, with respect to the best performing codec. In this paper, we compare the performance of emerging VVC and AV1 codecs, along with popular HEVC implementations, namely the HEVC Test Model (HM) and x265, as well as earlier, VP9 codec, and investigate their suitability for medical applications. To the best of our knowledge, this is the first performance comparison of emerging VVC and AV1 video codecs for use in the healthcare domain. Experimental evaluation based on three datasets (ultrasound, emergency scenery, and general-purpose videos) demonstrate that VVC outperforms all rival codecs while AV1 achieves better compression efficiency than HEVC in all cases but low-resolution (560\times448 @40Hz) ultrasound videos of the common carotid artery. Furthermore, the use of video despeckling prior to ultrasound video compression can provide significant bitrate savings.
The aim of this case report was to evaluate the feasibility, efficacy, and safety of nerve-sparing laparoscopic sacrocolpopexy (SCP) performed with a minimally invasive approach by using 2.9-mm ...Senhance ® surgical robotic system (Senhance ® , TRANSENTERIX Inc., USA). A 60-year-old Caucasian woman with symptomatic pelvic organ prolapse-Q (POP-Q) Aa: 2, Ba: 3, C: +4, Bp:2, Ap: 2, TVL:10 underwent subtotal hysterectomy with bilateral salpingo-oophorectomy, with nerve-sparing SCP performed using the Senhance surgical robotic system.. The urogynaecological assessment on the day of discharge and at the 3 month follow-up showed surgical anatomic success (<2 POP-Q stage). The patient was fully satisfied with the cosmetic result. This is the first case of SCP performed with this innovative system. SCP using “Senhance ®” is a feasible and effective approach with good results in terms of operative time, cosmesis, postoperative pain and length of hospitalisation.
The development of technologies that can generate Deepfake videos is expanding rapidly. These videos are easily synthesized without leaving obvious traces of manipulation. Though forensically ...detection in high-definition video datasets has achieved remarkable results, the forensics of compressed videos is worth further exploring. In fact, compressed videos are common in social networks, such as videos from Instagram, Wechat, and Tiktok. Therefore, how to identify compressed Deepfake videos becomes a fundamental issue. In this paper, we propose a two-stream method by analyzing the frame-level and temporality-level of compressed Deepfake videos. Since the video compression brings lots of redundant information to frames, the proposed frame-level stream gradually prunes the network to prevent the model from fitting the compression noise. Aiming at the problem that the temporal consistency in Deepfake videos might be ignored, we apply a temporality-level stream to extract temporal correlation features. When combined with scores from the two streams, our proposed method performs better than the state-of-the-art methods in compressed Deepfake videos detection.
Over-the-top mobile adaptive video streaming is invariably influenced by volatile network conditions, which can cause playback interruptions (stalling or rebuffering events) and bitrate fluctuations, ...thereby impairing users' quality of experience (QoE). Video quality assessment models that can accurately predict users' QoE under such volatile network conditions are rapidly gaining attention, since these methods could enable more efficient design of quality control protocols for media-driven services such as YouTube, Amazon, Netflix, and many others. However, the development of improved QoE prediction models requires data sets of videos afflicted with diverse stalling events that have been labeled with ground-truth subjective opinion scores. Toward this end, we have created a new mobile video quality database that we call LIVE Mobile Stall Video Database-II. Our database contains a total of 174 videos afflicted with distortions caused by 26 different stalling patterns. We describe the way we simulated the diverse stalling events to create a corpus of distorted videos, and we detail the human study we conducted to obtain continuous-time subjective scores from 54 subjects. We also present the outcomes of our comprehensive analysis of the impact of several factors that influence subjective QoE, and report the performance of existing QoE-prediction models on our data set. We are making the database (videos, subjective data, and video metadata) publicly available in order to help the advance state-of-the-art research on user-centric mobile network planning and management. The database may be accessed at http://live.ece.utexas.edu/research/LIVEStallStudy/liveMobile.html .
Significant improvements in video compression capability have been demonstrated with the introduction of the H.264/MPEG-4 advanced video coding (AVC) standard. Since developing this standard, the ...Joint Video Team of the ITU-T Video Coding Experts Group (VCEG) and the ISO/IEC Moving Picture Experts Group (MPEG) has also standardized an extension of that technology that is referred to as multiview video coding (MVC). MVC provides a compact representation for multiple views of a video scene, such as multiple synchronized video cameras. Stereo-paired video for 3-D viewing is an important special case of MVC. The standard enables inter-view prediction to improve compression capability, as well as supporting ordinary temporal and spatial prediction. It also supports backward compatibility with existing legacy systems by structuring the MVC bitstream to include a compatible "base view." Each other view is encoded at the same picture resolution as the base view. In recognition of its high-quality encoding capability and support for backward compatibility, the stereo high profile of the MVC extension was selected by the Blu-Ray Disc Association as the coding format for 3-D video with high-definition resolution. This paper provides an overview of the algorithmic design used for extending H.264/MPEG-4 AVC towards MVC. The basic approach of MVC for enabling inter-view prediction and view scalability in the context of H.264/MPEG-4 AVC is reviewed. Related supplemental enhancement information (SEI) metadata is also described. Various "frame compatible" approaches for support of stereo-view video as an alternative to MVC are also discussed. A summary of the coding performance achieved by MVC for both stereo- and multiview video is also provided. Future directions and challenges related to 3-D video are also briefly discussed.
High Efficiency Video Coding and Other Emerging Standards provides an overview of high efficiency video coding (HEVC) and all its extensions and profiles. There are nearly 300 projects and problems ...included, and about 400 references related to HEVC alone. Next generation video coding (NGVC) beyond HEVC is also described. Other video coding standards such as AVS2, DAALA, THOR, VP9 (Google), DIRAC, VC1, and AV1 are addressed, and image coding standards such as JPEG, JPEG-LS, JPEG2000, JPEG XR, JPEG XS, JPEG XT and JPEG-Pleno are also listed.Understanding of these standards and their implementation is facilitated by overview papers, standards documents, reference software, software manuals, test sequences, source codes, tutorials, keynote speakers, panel discussions, reflector and ftp/web sites - all in the public domain. Access to these categories is also provided.