scispace - formally typeset
Search or ask a question

Showing papers by "Mitre Corporation published in 2007"


Journal ArticleDOI
25 Jan 2007
TL;DR: eTuner, an approach to automatically tune schema matching systems, is described, which produced tuned matching systems that achieve higher accuracy than using the systems with currently possible tuning methods.
Abstract: Most recent schema matching systems assemble multiple components, each employing a particular matching technique. The domain user mustthen tune the system: select the right component to be executed and correctly adjust their numerous “knobs” (e.g., thresholds, formula coefficients). Tuning is skill and time intensive, but (as we show) without it the matching accuracy is significantly inferior. We describe eTuner, an approach to automatically tune schema matching systems. Given a schema S, we match S against synthetic schemas, for which the ground truth mapping is known, and find a tuning that demonstrably improves the performance of matching S against real schemas. To efficiently search the huge space of tuning configurations, eTuner works sequentially, starting with tuning the lowest level components. To increase the applicability of eTuner, we develop methods to tune a broad range of matching components. While the tuning process is completely automatic, eTuner can also exploit user assistance (whenever available) to further improve the tuning quality. We employed eTuner to tune four recently developed matching systems on several real-world domains. The results show that eTuner produced tuned matching systems that achieve higher accuracy than using the systems with currently possible tuning methods.

155 citations


Journal ArticleDOI
TL;DR: This paper describes a successful approach to de-identification that was developed to participate in a recent AMIA-sponsored challenge evaluation, and developed a method for tuning the balance of recall vs. precision in the Carafe system.

134 citations


Proceedings ArticleDOI
17 Apr 2007
TL;DR: In this article, a matched filter coherently integrates the radar data even though the target scatterers move through many range resolution cells during the coherent integration time, which can produce simultaneous high range and high Doppler resolution matched filter outputs.
Abstract: We have developed and demonstrated the Keystone format to simultaneously remove linear range migration for all targets regardless of their velocities. Higher order motion and under sampling foldover can be removed by hypotheses. The authors present an approach to radar matched filtering which can produce simultaneous high range and high Doppler resolution matched filter outputs. The new matched filter coherently integrates the radar data even though the target scatterers move through many range resolution cells during the coherent integration time.

124 citations


Journal ArticleDOI
30 Oct 2007-Vine
TL;DR: In this article, the authors explore the knowledge sharing factors from non-executive employees' perspective and find that the willingness of employees to share knowledge is the dependent variable for knowledge sharing.
Abstract: Purpose – Research has shown that the current knowledge management (KM) practices are developed from the standpoint of the management and do not put enough emphasis on knowledge sharing from the non‐executive employees' perspective However, it is important for organizations to understand – from the perspective of employees – the factors that motivate employees to share knowledge for successful implementation of any KM program In this exploratory study, willingness of employees to share knowledge is the dependent variable The purpose of this study is to explore the knowledge sharing factors from the employees' perspectiveDesign/methodology/approach – Using survey methodology, two large IT service and consulting organizations were included in the study to examine cultural, technological, motivational and organizational factors, which influence knowledge sharing within an organization from the perspective of non‐executive employeesFindings – The study results showed that issues related to availability a

115 citations


Book ChapterDOI
05 Sep 2007
TL;DR: In this article, the authors developed an approach for detecting insiders who operate outside the scope of their duties and thus violate need-to-know, based on information from public cases, consultation with domain experts, and analysis of a massive collection of information-use events and contextual information.
Abstract: Malicious insiders do great harm and avoid detection by using their legitimate privileges to steal information that is often outside the scope of their duties. Based on information from public cases, consultation with domain experts, and analysis of a massive collection of information-use events and contextual information, we developed an approach for detecting insiders who operate outside the scope of their duties and thus violate need-to-know. Based on the approach, we built and evaluated elicit, a system designed to help analysts investigate insider threats. Empirical results suggest that, for a specified decision threshold of .5, elicit achieves a detection rate of .84 and a false-positive rate of .015, flagging per day only 23 users of 1, 548 for further scrutiny. It achieved an area under an roc curve of .92.

106 citations


Journal ArticleDOI
Daniel Loehr1
01 Jan 2007-Gesture
TL;DR: In this article, the authors investigated the relationship between gesture and speech and found a rich rhythmic relationship between the hands, head, and voice, with a common tempo of around a third of a second, perhaps to synchronize inter-articulator meeting points.
Abstract: This article investigates the rhythmic relationship between gesture and speech. Four subjects were filmed in natural conversations with friends. From the resulting videos, several thousand time-stamped annotations pertaining to rhythm were manually recorded in a digital annotation tool, and exported for statistical analysis. They revealed a rich rhythmic relationship between the hands, head, and voice. Each articulator produced pikes (a general term for short, distinctive expressions, regardless of the modality) in complex synchrony with other articulators. Even eyeblinks were synchronized, with eyelids held closed until reopening on the rhythmic beat, akin to a pre-stroke hold before a gestural stroke. Average tempos similar to previously reported natural human tempos — e.g. Fraisse’s (1982) 600 ms figure — were found in hands, head, and speech, although hands tended to move most quickly and speech most slowly. All three also shared a common tempo of around a third of a second, perhaps to synchronize inter-articulator meeting points. These findings lend empirical weight to earlier observations of a rhythmic relationship between gesture and speech, providing support for the theory of a common cognitive origin of the two modalities.

105 citations


Book ChapterDOI
24 Mar 2007
TL;DR: A method for enumerating all essentially different executions possible for a cryptographic protocol, called the shapes of the protocol, and cpsa, the Cryptographic Protocol Shape Analyzer, implements the method.
Abstract: We describe a method for enumerating all essentially different executions possible for a cryptographic protocol. We call them the shapes of the protocol. Naturally occurring protocols have only finitely many, indeed very few shapes. Authentication and secrecy properties are easy to determine from them, as are attacks. cpsa, our Cryptographic Protocol Shape Analyzer, implements the method. In searching for shapes, cpsa starts with some initial behavior, and discovers what shapes are compatible with it. Normally, the initial behavior is the point of view of one participant. The analysis reveals what the other principals must have done, given this participant's view.

103 citations


Journal ArticleDOI
01 Sep 2007
TL;DR: The use of the Virtual Address Descriptor tree structure in Windows memory dumps to help guide forensic analysis of Windows memory is described and its value in breaking up physical memory into more manageable and semantically meaningful units is shown.
Abstract: This paper describes the use of the Virtual Address Descriptor (VAD) tree structure in Windows memory dumps to help guide forensic analysis of Windows memory. We describe how to locate and parse the structure, and show its value in breaking up physical memory into more manageable and semantically meaningful units than can be obtained by simply walking the page directory for the process. Several tools to display information about the VAD tree and dump the memory regions it describes will also be presented.

86 citations


Journal ArticleDOI
TL;DR: In this article, the behavior of a matched electrically small antenna that exhibits two impedance resonances within its defined voltage standing wave ratio (VSWR) bandwidth was examined, and the exact quality factor (Q), computed indirectly from the antenna's input reactance and far field, was compared to twice the inverse of the matched half-power VSWR bandwidth.
Abstract: We examine the behavior of a matched electrically small antenna that exhibits two impedance resonances within its defined voltage standing wave ratio (VSWR) bandwidth. The exact quality factor (Q), computed indirectly from the antenna's input reactance and far field, is compared to twice the inverse of the matched half-power VSWR bandwidth (Q BW) and to an approximate quality factor (Qz) determined from the frequency derivative of the antenna's input impedance. The well-established approximate equalities Q ap Q BW ap Q z for antennas exhibiting an isolated single impedance resonance within their operating band are shown to become highly inaccurate for an electrically small antenna exhibiting two closely spaced impedance resonances.

77 citations


Journal ArticleDOI
TL;DR: The authors present tradeoffs between the use of matched versus mismatched filters, loss in signal-to-noise ratio (SNR), time sidelobe levels, and Doppler intolerance for random and cyclic codes.
Abstract: This correspondence focuses on the use of phase-coded (PC) waveforms for radar. The authors present tradeoffs between the use of matched versus mismatched filters, loss in signal-to-noise ratio (SNR), time sidelobe levels, and Doppler intolerance for random and cyclic codes. A multiple hypothesis approach is introduced for countering intrapulse Doppler intolerance

67 citations


Journal ArticleDOI
TL;DR: The study methodology, the competitors’ systems and performance, and ways to improve human-robot interaction in urban search and rescue (USAR) as well as other remote robot operations are discussed.
Abstract: This paper presents results from three years of studying human-robot interaction in the context of the AAAI Robot Rescue Competition. We discuss our study methodology, the competitors' systems and performance, and suggest ways to improve human-robot interaction in urban search and rescue (USAR) as well as other remote robot operations.

Proceedings ArticleDOI
10 Mar 2007
TL;DR: It is shown that a map-centric interface is more effective in providing good location and status awareness while a video-centric interfaces are more effectiveIn providing good surroundings and activities awareness.
Abstract: Good situation awareness (SA) is especially necessary when robots and their operators are not collocated, such as in urban search and rescue (USAR). This paper compares how SA is attained in two systems: one that has an emphasis on video and another that has an emphasis on a three-dimensional map. We performed a within-subjects study with eight USAR domain experts. To analyze the utterances made by the participants, we developed a SA analysis technique, called LASSO, which includes five awareness categories: location, activities, surroundings, status, and overall mission. Using our analysis technique, we show that a map-centric interface is more effective in providing good location and status awareness while a video-centric interface is more effective in providing good surroundings and activities awareness.

Proceedings ArticleDOI
09 Jun 2007
TL;DR: In this paper, several top loaded monopole antennas leading to the Goubau multi-element configuration are investigated and their input impedance, bandwidth, and radiation pattern characteristics obtained using full-wave simulators, are discussed in detail.
Abstract: Several top loaded monopole antennas leading to the Goubau multi element configuration are investigated. Their input impedance, bandwidth, and radiation pattern characteristics obtained using full-wave simulators, are discussed in detail.

Proceedings ArticleDOI
08 May 2007
TL;DR: The Web Mashup Scripting Language (WMSL) enables an end-user working from his browser, e.g. not needing any other infrastructure, to quickly write mashups that integrate any two, or more, web services on the Web.
Abstract: The Web Mashup Scripting Language (WMSL) enables an end-user (you) working from his browser, e.g. not needing any other infrastructure, to quickly write mashups that integrate any two, or more, web services on the Web. The end-user accomplishes this by writing a web page that combines HTML, metadata in the form of mapping relations, and small piece of code, or script. The mapping relations enable not only the discovery and retrieval of the WMSL pages, but also affect a new programming paradigm that abstracts many programming complexities from the script writer. Furthermore, the WMSL Web pages or scripts that disparate end-users (you) write, can be harvested by Crawlers to automatically generate the concepts needed to build lightweight ontologies containing local semantics of a web service and its data model, to extend context ontologies or middle ontologies, and to develop links, or mappings, between these ontologies. This enables an open-source model of building ontologies based on the WMSL Web page or scripts that end users (you) write.

Proceedings ArticleDOI
Laurie Damianos1, D. Cuomo1, J. Griffith1, D.M. Hirst1, J. Smallwood1 
03 Jan 2007
TL;DR: It is hypothesize that social bookmarking would be useful in this environment for resource management, information sharing and discovery, expert finding, and social networking, and the challenges and plans for future development and integration into the enterprise.
Abstract: This paper describes an ongoing technology investigation to assess the value and utility of social bookmarking on a corporate intranet. We hypothesize that social bookmarking would be useful in this environment for resource management, information sharing and discovery, expert finding, and social networking. We discuss features of the prototype system deployed and early analysis of findings on adoption, usage, and social influences. We conclude with our challenges and plans for future development and integration into the enterprise

Proceedings ArticleDOI
09 Nov 2007
TL;DR: This work correctly identifies the content portion of web-pages anywhere from 80-97% of the time depending on experimental factors such as ensuring the absence of duplicate documents and application of the model against unseen sources.
Abstract: Identifying which parts of a Web-page contain target content (e.g., the portion of an online news page that contains the actual article) is a significant problem that must be addressed for many Web-based applications. Most approaches to this problem involve crafting hand-tailored rules or scripts to extract the content, customized separately for particular Web sites. Besides requiring considerable time and effort to implement, hand-built extraction routines are brittle: they fail to properly extract content in some cases and break when the structure of a site's Web-pages changes. In this work we treat the problem of identifying content as a sequence labeling problem, a common problem structure in machine learning and natural language processing. Using a Conditional Random Field sequence labeling model, we correctly identify the content portion of web-pages anywhere from 80-97% of the time depending on experimental factors such as ensuring the absence of duplicate documents and application of the model against unseen sources.

Journal ArticleDOI
TL;DR: This paper demonstrates via simulation that the nanodevices and nanofabrication techniques developed recently for general-purpose nanocomputers also might be applied with substantial benefit to implement less complex nanocircuits targeted at specific applications.
Abstract: Designs and simulation results are given for two small, special-purpose nanoelectronic circuits. The area of special-purpose nanoelectronics has not been given much consideration previously, though much effort has been devoted to the development of general-purpose nanoelectronic systems, i.e., nanocomputers. This paper demonstrates via simulation that the nanodevices and nanofabrication techniques developed recently for general-purpose nanocomputers also might be applied with substantial benefit to implement less complex nanocircuits targeted at specific applications. Nanocircuits considered here are a digital controller for the leg motion on an autonomous millimeter-scale robot and an analog nanocircuit for amplification of signals in a tiny optoelectronic sensor or receiver. Simulations of both nanocircuit designs show significant improvement over microelectronic designs in metrics such as footprint area and power consumption. These improvements are obtained from designs employing nanodevices and nanofabrication techniques that already have been demonstrated experimentally. Thus, the results presented here suggest that such improvements might be realized in the near term for important, special-purpose applications.

Proceedings ArticleDOI
25 Jun 2007
TL;DR: This paper defines interoperability; discusses why the lack of medical device interoperability is a problem, and the scope of the problem; and numerous benefits to medical devices interoperability are outlined.
Abstract: As we know, interoperability is an almost nonexistent feature of medical devices. This paper defines interoperability; discusses why the lack of medical device interoperability is a problem, and the scope of the problem. Possible reasons for the lack of medical device interoperability are explored. Numerous benefits to medical device interoperability are outlined and possible risks of maintaining the status quo are predicted.

Journal ArticleDOI
TL;DR: The rationale and construction of the L1C spreading codes families were created from extensive searches with each search requiring its own fine-tuned techniques and search criteria.
Abstract: We describe the rationale and construction of the family of L1C spreading codes and the associated family of overlay codes. The length-10230 spreading codes comprise 210 pilot/data sequence pairs. The codes are perfectly balanced and exhibit good even and odd auto- and crosscorrelation. They are constructed using prime-length Weil codes: selected length-10223 Weil-codes were padded with a fixed 7-bit pad. A search over all Weil codes and insertion points yielded a set of 739 codes from which the final set of 420 codes was chosen. Two separate families of codes were optimized based on modulation scheme (BOC(1,1) and MBOC). The length-1800 overlay codes are truncated linear feedback shift register sequences of length 2047, including both m-sequences and Gold sequences. Besides full period even auto- and cross-correlation, the overlay code search considered the correlation when a small subsequence (100 or 200 symbols) is correlated against the full code.

Proceedings ArticleDOI
01 Oct 2007
TL;DR: A simulation evaluation of MANET protocol performance for an airborne environment is presented, with the intent to identify a routing protocol that can best deal with the dynamics of an airborne network.
Abstract: Communications among mobile, tactical nodes presents a major military challenge. The use of MANET (Mobile Ad Hoc Network) protocols provides a possible solution for military nodes, including those in an airborne network. However MANET research has primarily focused on ground-based studies, using vehicular speeds and in many cases random mobility patterns. Nodes of an airborne network travel at speeds significantly faster than ground vehicles, and fly in coordinated paths not modeled by random mobility. In addition, the quality of the radio links for airborne nodes varies with time, due to interference, range, or antenna occlusion when banking. These characteristics make it impossible to extrapolate existing MANET research results to the airborne network. In this paper we present a simulation evaluation of MANET protocol performance for an airborne environment, with the intent to identify a routing protocol that can best deal with the dynamics of an airborne network. A scenario involving widebody aircraft trajectories was modeled in OPNET. Intermittent link outages due to aircraft banking were modeled by use of a notional radio link, antenna model, and modified OPNET source code that reflects positional antenna gain, including antenna occlusion when an aircraft banks. Within this scenario environment, four MANET protocols (AODV, TORA, OLSR, OSPFv3-MANET) were run on the airborne nodes with metric collection of protocol overhead, packet delivery ratio, and packet delay. Simulation results and analysis of the protocol performance for an airborne network are presented here. Additional issues and future areas of research are also identified.

Journal ArticleDOI
TL;DR: This paper describes two steps in the evolution of human‐robot interaction designs developed by the University of Massachusetts Lowell (UML) and the Idaho National Laboratory to support urban search and rescue tasks and suggests that performance is better with the new interaction techniques.
Abstract: This paper describes two steps in the evolution of human-robot interaction designs developed by the University of Massachusetts Lowell (UML) and the Idaho National Laboratory to support urban search and rescue tasks. Usability tests were conducted to compare the two interfaces, one of which emphasized three-dimensional mapping while the other design emphasized the video feed. We found that participants desired a combination of the interface design approaches. As a result, the UML system was changed to augment its heavy emphasis on video with a map view of the area immediately around the robot. The changes were tested in a follow-up user study and the results from that experiment suggest that performance, as measured by the number of collisions with objects in the environment and time on task, is better with the new interaction techniques. Throughout the paper, we describe how we applied human-computer interaction principles and techniques to benefit the evolution of the human-robot interaction designs. While the design work is situated in the urban search and rescue domain, the results can be generalized to domains that involve other search or monitoring tasks using remotely located robots. © 2007 Wiley Periodicals, Inc.

Journal ArticleDOI
TL;DR: A new approach towards identification of tandem repeats in DNA sequences is proposed, a refinement of previously considered method, based on the complex periodicity transform, obtained by mapping of DNA symbols to pure quaternions.
Abstract: Motivation: One of the main tasks of DNA sequence analysis is identification of repetitive patterns. DNA symbol repetitions play a key role in a number of applications, including prediction of gene and exon locations, identification of diseases, reconstruction of human evolutionary history and DNA forensics. Results: A new approach towards identification of tandem repeats in DNA sequences is proposed. The approach is a refinement of previously considered method, based on the complex periodicity transform. The refinement is obtained, among others, by mapping of DNA symbols to pure quaternions. This mapping results in an enhanced, symbol-balanced sensitivity of the transform to DNA patterns, and an unambiguous threshold selection criterion. Computational efficiency of the transform is further improved, and coupling of the computation with the period value is removed, thereby facilitating parallel implementation of the algorithm. Additionally, a post-processing stage is inserted into the algorithm, enabling unambiguous display of results in a convenient graphical format. Comparison of the quaternionic periodicity transform with two well-known pattern detection techniques shows that the new approach is competitive with these two techniques in detection of exact and approximate repeats. Supplementary information: Supplementary data are available at Bioinformatics online.

Journal ArticleDOI
TL;DR: Researchers aim to have this simulation assist air traffic control (ATC) in solving various air traffic congestion contingencies by showing cost-benefit analyses of various decisions involved.
Abstract: En route airspace congestion, often due to convective weather, causes system-wide delays and disruption in the US National Airspace System (NAS). Present-day methods for managing congestion are mos...

Posted Content
TL;DR: This work addresses several of the concerns expressed in previous work, such as negation, complementary classes, disjunctive heads, and cardinality, and it discusses alternative approaches for dealing with inconsistencies in the knowledge base.
Abstract: To appear in Theory and Practice of Logic Programming (TPLP), 2008. We are researching the interaction between the rule and the ontology layers of the Semantic Web, by comparing two options: 1) using OWL and its rule extension SWRL to develop an integrated ontology/rule language, and 2) layering rules on top of an ontology with RuleML and OWL. Toward this end, we are developing the SWORIER system, which enables efficient automated reasoning on ontologies and rules, by translating all of them into Prolog and adding a set of general rules that properly capture the semantics of OWL. We have also enabled the user to make dynamic changes on the fly, at run time. This work addresses several of the concerns expressed in previous work, such as negation, complementary classes, disjunctive heads, and cardinality, and it discusses alternative approaches for dealing with inconsistencies in the knowledge base. In addition, for efficiency, we implemented techniques called extensionalization, avoiding reanalysis, and code minimization.

BookDOI
01 Jan 2007
TL;DR: In this article, an agent-based approach to user-initiated Semantic Service Interconnection is presented, along with a Lightweight Agent Fabric for Service Autonomy and Role-Based Support Mechanism for Service Description and Discovery.
Abstract: Executing Semantic Web Services with a Context-Aware Service Execution Agent.- An Effective Strategy for the Flexible Provisioning of Service Workflows.- Using Goals for Flexible Service Orchestration.- An Agent-Based Approach to User-Initiated Semantic Service Interconnection.- A Lightweight Agent Fabric for Service Autonomy.- Semantic Service Composition in Service-Oriented Multiagent Systems: A Filtering Approach.- Towards a Mapping from BPMN to Agents.- Associated Topic Extraction for Consumer Generated Media Analysis.- An MAS Infrastructure for Implementing SWSA Based Semantic Services.- A Role-Based Support Mechanism for Service Description and Discovery.- WS2JADE: Integrating Web Service with Jade Agents.- Z-Based Agents for Service Oriented Computing.

Proceedings ArticleDOI
10 Mar 2007
TL;DR: A case study in the urban search-and-rescue domain is provided and rationale for selecting different types of GOMS modeling techniques to help the analyst model human-robot interfaces is provided.
Abstract: A formal interaction modeling technique known as Goals, Operators, Methods, and Selection rules (GOMS) is well-established in human-computer interaction as a cost-effective way of evaluating designs without the participation of end users. This paper explores the use of GOMS for evaluating human-robot interaction. We provide a case study in the urban search-and-rescue domain and raise issues for developing GOMS models that have not been previously addressed. Further, we provide rationale for selecting different types of GOMS modeling techniques to help the analyst model human-robot interfaces.

24 Jun 2007
TL;DR: A design methodology that adds value to engineering systems by considering flexibility at an early conceptual stage and provides screening tools to find areas where flexibility can be incorporated at the engineering, operational, and management decision levels is introduced.
Abstract: Designers and managers of new investments in engineering systems look for ways to add value to their programs. One fundamental way to do this is by taking advantage of uncertainty. Although uncertainty is usually seen as negative in most investment projects, it can also increase performance if flexibility is incorporated into the system to capture upside opportunities, and reduce losses in case of downside events. This paper introduces a design methodology that adds value to engineering systems by considering flexibility at an early conceptual stage. It provides screening tools to find areas where flexibility can be incorporated at the engineering, operational, and management decision levels. In engineering and operations, technical modifications need to be done within the system to acquire the flexibility exercisable by managers. One example is the ability to expand or contract product output as demand fluctuates. At the management decision level, no explicit modification is needed, such as the ability to abandon the project altogether. The methodology incorporates screening tools based both on qualitative historical studies (GPS, B-52, etc.) and quantitative Design Structure Matrices representing the engineering system (Bartolomei et al. 2006; Kalligeros 2006; Kalligeros, de Neufville 2006). The design process also provides a set of quantitative tools to assess the financial value of flexibility based on Real Options Analysis and simulation models (de Neufville et al. 2006; Kalligeros 2006; Kalligeros, de Neufville 2006). These give managers and designers discriminating tools with which to choose the most valuable flexibilities to implement in the engineering system. The methodology represents a practical procedure for understanding where flexibility can be found and incorporated into all areas of engineering system design. We present a case application to the design of a new hydrogen production and storage system: Fusion Island (Cardin 2006; Nuttall, et al. 2005).


15 May 2007
TL;DR: This special issue of TAL looks at the fundamental principles underlying evaluation in natural language processing and adopts a global point of view that goes beyond the horizon of a single evaluation campaign or a particular protocol.
Abstract: In this special issue of TAL, we look at the fundamental principles underlying evaluation in natural language processing. We adopt a global point of view that goes beyond the horizon of a single evaluation campaign or a particular protocol. After a brief review of history and terminology, we will address the topic of a gold standard for natural language processing, of annotation quality, of the amount of data, of the difference between technology evaluation and usage evaluation, of dialog systems, and of standards, before concluding with a short discussion of the articles in this special issue and some prospective remarks

Journal ArticleDOI
12 Dec 2007-PLOS ONE
TL;DR: Systematic characterization of the spatio-temporal distribution of TB cases can widely benefit real time surveillance and guide public health investigations of TB outbreaks as to what level of spatial resolution results in improved detection sensitivity and timeliness.
Abstract: Background. San Francisco has the highest rate of tuberculosis (TB) in the U.S. with recurrent outbreaks among the homeless and marginally housed. It has been shown for syndromic data that when exact geographic coordinates of individual patients are used as the spatial base for outbreak detection, higher detection rates and accuracy are achieved compared to when data are aggregated into administrative regions such as zip codes and census tracts. We examine the effect of varying the spatial resolution in the TB data within the San Francisco homeless population on detection sensitivity, timeliness, and the amount of historical data needed to achieve better performance measures. Methods and Findings. We apply a variation of space-time permutation scan statistic to the TB data in which a patient’s location is either represented by its exact coordinates or by the centroid of its census tract. We show that the detection sensitivity and timeliness of the method generally improve when exact locations are used to identify real TB outbreaks. When outbreaks are simulated, while the detection timeliness is consistently improved when exact coordinates are used, the detection sensitivity varies depending on the size of the spatial scanning window and the number of tracts in which cases are simulated. Finally, we show that when exact locations are used, smaller amount of historical data is required for training the model. Conclusion. Systematic characterization of the spatio-temporal distribution of TB cases can widely benefit real time surveillance and guide public health investigations of TB outbreaks as to what level of spatial resolution results in improved detection sensitivity and timeliness. Trading higher spatial resolution for better performance is ultimately a tradeoff between maintaining patient confidentiality and improving public health when sharing data. Understanding such tradeoffs is critical to managing the complex interplay between public policy and public health. This study is a step forward in this direction.