Papers made digestable
Our architecture simplifies the obstacle-perception
problem to that of place-dependent change detection. While we use the method with VT&R, it
can be generalized to suit arbitrary path-following applications.
Visual Teach and Repeat 3 (VT&R3), a generalization of stereo VT&R, achieves
long-term autonomous path-following using topometric mapping and localization
from a single rich sensor stream. In this paper, we improve the capabilities of
a LiDAR implementation of VT&R3 to reliably detect and avoid obstacles in
changing environments. Our architecture simplifies the obstacle-perception
problem to that of place-dependent change detection. We then extend the
behaviour of generic sample-based motion planners to better suit the
teach-and-repeat problem structure by introducing a new edge-cost metric paired
with a curvilinear planning space. The resulting planner generates naturally
smooth paths that avoid local obstacles while minimizing lateral path deviation
to best exploit prior terrain knowledge. While we use the method with VT&R, it
can be generalized to suit arbitrary path-following applications. Experimental
results from online run-time analysis, unit testing, and qualitative
experiments on a differential drive robot show the promise of the technique for
reliable long-term autonomous operation in complex unstructured environments.
Authors: Jordy Sehn, Yuchen Wu, Timothy D. Barfoot.
The statistical and design considerations that pertain to
dose optimization are discussed. The sample size savings range from 16.6% to 27.3%,
depending on the design and scenario, with a mean savings of 22.1%.
The traditional more-is-better dose selection paradigm, developed based on
cytotoxic chemotherapeutics, is often problematic When applied to the
development of novel molecularly targeted agents (e.g., kinase inhibitors,
monoclonal antibodies, and antibody-drug conjugates). The US Food and Drug
Administration (FDA) initiated Project Optimus to reform the dose optimization
and dose selection paradigm in oncology drug development and call for more
attention to benefit-risk consideration.
We systematically investigated the operating characteristics of the seamless
phase 2-3 design as a strategy for dose optimization, where in stage 1
(corresponding to phase 2) patients are randomized to multiple doses, with or
without a control; and in stage 2 (corresponding to phase 3) the efficacy of
the selected optimal dose is evaluated with a randomized concurrent control or
historical control. Depending on whether the concurrent control is included and
the type of endpoints used in stages 1 and 2, we describe four types of
seamless phase 2-3 dose-optimization designs, which are suitable for different
clinical settings. The statistical and design considerations that pertain to
dose optimization are discussed. Simulation shows that dose optimization phase
2-3 designs are able to control the familywise type I error rates and yield
appropriate statistical power with substantially smaller sample size than the
conventional approach. The sample size savings range from 16.6% to 27.3%,
depending on the design and scenario, with a mean savings of 22.1%. Due to the
interim dose selection, the phase 2-3 dose-optimization design is logistically
and operationally more challenging, and should be carefully planned and
implemented to ensure trial integrity.
Authors: Liyun Jiang, Ying Yuan.
We significantly improve performance using properties of the posterior
in our active learning scheme and for the definition of the GP prior. In
particular we account for the expected dynamical range of the posterior in
different dimensionalities. We test our model against a number of synthetic and
cosmological examples.
We present the GPry algorithm for fast Bayesian inference of general
(non-Gaussian) posteriors with a moderate number of parameters. GPry does not
need any pre-training, special hardware such as GPUs, and is intended as a
drop-in replacement for traditional Monte Carlo methods for Bayesian inference.
Our algorithm is based on generating a Gaussian Process surrogate model of the
log-posterior, aided by a Support Vector Machine classifier that excludes
extreme or non-finite values. An active learning scheme allows us to reduce the
number of required posterior evaluations by two orders of magnitude compared to
traditional Monte Carlo inference. Our algorithm allows for parallel
evaluations of the posterior at optimal locations, further reducing wall-clock
times. We significantly improve performance using properties of the posterior
in our active learning scheme and for the definition of the GP prior. In
particular we account for the expected dynamical range of the posterior in
different dimensionalities. We test our model against a number of synthetic and
cosmological examples. GPry outperforms traditional Monte Carlo methods when
the evaluation time of the likelihood (or the calculation of theoretical
observables) is of the order of seconds; for evaluation times of over a minute
it can perform inference in days that would take months using traditional
methods. GPry is distributed as an open source Python package (pip install
gpry) and can also be found at https://github.com/jonaselgammal/GPry.
Authors: Jonas El Gammal, Nils Schöneberg, Jesús Torrado, Christian Fidler.
We consider the fundamental scheduling problem of minimizing the sum of
weighted completion times on a single machine in the non-clairvoyant setting. However, to the best of our knowledge, this concept has never been considered
for the total completion time objective in the non-clairvoyant model. This implies
a performance guarantee of $(1+3\sqrt{3})\approx 6.197$ for the deterministic
algorithm and of $\approx 3.032$ for the randomized version.
We consider the fundamental scheduling problem of minimizing the sum of
weighted completion times on a single machine in the non-clairvoyant setting.
While no non-preemptive algorithm is constant competitive, Motwani, Phillips,
and Torng (SODA '93) proved that the simple preemptive round robin procedure is
$2$-competitive and that no better competitive ratio is possible, initiating a
long line of research focused on preemptive algorithms for generalized variants
of the problem. As an alternative model, Shmoys, Wein, and Williamson (FOCS
'91) introduced kill-and-restart schedules, where running jobs may be killed
and restarted from scratch later, and analyzed then for the makespan objective.
However, to the best of our knowledge, this concept has never been considered
for the total completion time objective in the non-clairvoyant model.
We contribute to both models: First we give for any $b > 1$ a tight analysis
for the natural $b$-scaling kill-and-restart strategy for scheduling jobs
without release dates, as well as for a randomized variant of it. This implies
a performance guarantee of $(1+3\sqrt{3})\approx 6.197$ for the deterministic
algorithm and of $\approx 3.032$ for the randomized version. Second, we show
that the preemptive Weighted Shortest Elapsed Time First (WSETF) rule is
$2$-competitive for jobs released in an online fashion over time, matching the
lower bound by Motwani et al. Using this result as well as the competitiveness
of round robin for multiple machines, we prove performance guarantees of
adaptions of the $b$-scaling algorithm to online release dates and unweighted
jobs on identical parallel machines.
Authors: Sven Jäger, Guillaume Sagnol, Daniel Schmidt genannt Waldschmidt, Philipp Warode.
Frozen pretrained models have become a viable alternative to the
pretraining-then-finetuning paradigm for transfer learning. With this work, we hope to
bring greater attention to this promising path of freezing pretrained image
models.
Frozen pretrained models have become a viable alternative to the
pretraining-then-finetuning paradigm for transfer learning. However, with
frozen models there are relatively few parameters available for adapting to
downstream tasks, which is problematic in computer vision where tasks vary
significantly in input/output format and the type of information that is of
value. In this paper, we present a study of frozen pretrained models when
applied to diverse and representative computer vision tasks, including object
detection, semantic segmentation and video action recognition. From this
empirical analysis, our work answers the questions of what pretraining task
fits best with this frozen setting, how to make the frozen setting more
flexible to various downstream tasks, and the effect of larger model sizes. We
additionally examine the upper bound of performance using a giant frozen
pretrained model with 3 billion parameters (SwinV2-G) and find that it reaches
competitive performance on a varied set of major benchmarks with only one
shared frozen base network: 60.0 box mAP and 52.2 mask mAP on COCO object
detection test-dev, 57.6 val mIoU on ADE20K semantic segmentation, and 81.7
top-1 accuracy on Kinetics-400 action recognition. With this work, we hope to
bring greater attention to this promising path of freezing pretrained image
models.
Authors: Yutong Lin, Ze Liu, Zheng Zhang, Han Hu, Nanning Zheng, Stephen Lin, Yue Cao.
Recent experimental advances have inspired the development of theoretical
tools to describe the non-equilibrium dynamics of quantum systems. We benchmark our findings by considering analytically
solvable limits and providing alternative derivations of known results.
Recent experimental advances have inspired the development of theoretical
tools to describe the non-equilibrium dynamics of quantum systems. Among them
an exact representation of quantum spin systems in terms of classical
stochastic processes has been proposed. Here we provide first steps towards the
extension of this stochastic approach to bosonic systems by considering the
one-dimensional quantum quartic oscillator. We show how to exactly parameterize
the time evolution of this prototypical model via the dynamics of a set of
classical variables. We interpret these variables as stochastic processes,
which allows us to propose a novel way to numerically simulate the time
evolution of the system. We benchmark our findings by considering analytically
solvable limits and providing alternative derivations of known results.
Authors: Gennaro Tucci, Stefano De Nicola, Sascha Wald, Andrea Gambassi.
We investigate the connection of the regulation of star formation and the cycling of baryons within and in and out of galaxies. We use idealized numerical simulations of Milky Way-mass galaxies, in which we systemically vary the galaxy morphology (bulge-to-total mass ratio) and stellar feedback strength (total eight setups with 80 simulations). We conclude that measurements of the temporal and spatial PSD of the SFH can provide constraints on the baryon cycle and the star formation process.
We investigate the connection of the regulation of star formation and the
cycling of baryons within and in and out of galaxies. We use idealized
numerical simulations of Milky Way-mass galaxies, in which we systemically vary
the galaxy morphology (bulge-to-total mass ratio) and stellar feedback strength
(total eight setups with 80 simulations). By following individual gas parcels
through the disk, spiral arms, and massive star-forming clumps, we quantify how
gas moves and oscillates through the different phases of the interstellar
medium (ISM) and forms stars. We show that the residence time of gas in the
dense ISM phase ($\tau_{\rm SF}$), the nature of spiral arms (strength,
number), and the clump properties (number, mass function, and young star
fraction) depend on both the galaxy morphology and stellar feedback. Based on
these results, we quantify signatures of the baryon cycle within galaxies using
the temporal and spatial power spectrum density (PSD) of the star formation
history (SFH). Stronger stellar feedback leads to more bursty star formation
while the correlation timescale of the SFH is longer, because stronger feedback
dissolves the dense, star-forming ISM phase, leading to a more homogeneous ISM
and a decrease in $\tau_{\rm SF}$. The bulge strength has a similar effect: the
deep gravitational potential in a bulge-dominant galaxy imposes a strong shear
force that effectively breaks apart gas clumps in the ISM; this subsequently
inhibits the fragmentation of cool gas and therefore the star formation in the
disk, leading to a decrease in the spatial power on scales of $\sim$ 1 kpc. We
conclude that measurements of the temporal and spatial PSD of the SFH can
provide constraints on the baryon cycle and the star formation process.
Authors: Eun-jin Shin, Sandro Tacchella, Ji-hoon Kim, Kartheik G. Iyer, Vadim A. Semenov.
(2009). A series of
useful complementary results is also given. The
issue of testing multiple restrictions on the loadings as well as building
joint confidence intervals for the factors is discussed.
This paper revisits and provides an alternative derivation of the asymptotic
results for the Principal Components estimator of a large approximate factor
model as considered in Stock and Watson (2002), Bai (2003), and Forni et al.
(2009). Results are derived under a minimal set of assumptions with a special
focus on the time series setting, which is usually considered in almost all
recent empirical applications. Hence, $n$ and $T$ are not treated
symmetrically, the former being the dimension of the considered vector of time
series, while the latter being the sample size and, therefore, being relevant
only for estimation purposes, but not when it comes to just studying the
properties of the model at a population level. As a consequence, following
Stock and Watson (2002) and Forni et al. (2009), estimation is based on the
classical $n \times n$ sample covariance matrix. As expected, all asymptotic
results we derive are equivalent to those stated in Bai (2003), where, however,
a $T\times T$ covariance matrix is considered as a starting point. A series of
useful complementary results is also given. In particular, we give some
alternative sets of primitive conditions for mean-squared consistency of the
sample covariance matrix of the factors, of the idiosyncratic components, and
of the observed time series. We also give more intuitive asymptotic expansions
for the estimators showing that PCA is equivalent to OLS as long as
$\sqrt{T}/n\to 0$ and $\sqrt{n}/T\to 0$, that is loadings are estimated in a
time series regression as if the factors were known, while factors are
estimated in a cross-sectional regression as if the loadings were known. The
issue of testing multiple restrictions on the loadings as well as building
joint confidence intervals for the factors is discussed.
Authors: Matteo Barigozzi.
We also show that these quadratic triple testing conditions can be relaxed to scalar testing conditions, quadratic offset Muckenhoupt conditions, and a quadratic weak boundedness property.
If T is a smooth Stein elliptic fractional singular integral, 1<p<infinity,
and sigma and omega are doubling measures, then the two weight L^{p} norm
inequality holds if and only if the quadratic triple testing conditions of
Hyt\"onen and Vuorinen hold. We also show that these quadratic triple testing
conditions can be relaxed to scalar testing conditions, quadratic offset
Muckenhoupt conditions, and a quadratic weak boundedness property.
Authors: Eric T. Sawyer, Brett D. Wick.
Theoretical methods that are accurate for both short-distance observables and
long-wavelength collective modes are still being developed for the Hubbard
model. The DMFT double occupancy determines the
spin and charge vertices through local spin and charge sum rules. The TPSC
self-energy is also improved by replacing its local part with the local DMFT
self-energy. We also find that the accuracy check
developed for TPSC is a good predictor of deviations from benchmarks. TPSC+DMFT
can be used in regimes where quantum Monte Carlo is inaccessible.
Theoretical methods that are accurate for both short-distance observables and
long-wavelength collective modes are still being developed for the Hubbard
model. Here, we benchmark against published diagrammatic quantum Monte Carlo
results an approach that combines local observables from dynamical mean-field
theory (DMFT) with the two-particle self-consistent theory (TPSC). This method
(TPSC+DMFT) is relevant for weak to intermediate interaction, satisfies the
local Pauli principle and allows us to compute a spin susceptibility that
satisfies the Mermin-Wagner theorem. The DMFT double occupancy determines the
spin and charge vertices through local spin and charge sum rules. The TPSC
self-energy is also improved by replacing its local part with the local DMFT
self-energy. With this method, we find improvements for both spin and charge
fluctuations and for the self-energy. We also find that the accuracy check
developed for TPSC is a good predictor of deviations from benchmarks. TPSC+DMFT
can be used in regimes where quantum Monte Carlo is inaccessible. In addition,
this method paves the way to multi-band generalizations of TPSC that could be
used in advanced electronic structure codes that include DMFT.
Authors: N. Martin, C. Gauvin-Ndiaye, A. -M. S. Tremblay.
Infinite-dimensional control systems with outputs are considered in the Hamiltonian formulation with generalized coordinates. Sufficient conditions for the convergence of the constructed observer are obtained on the basis of the invariance principle. The estimation error decay is illustrated with numerical simulations of finite-dimensional approximations of the observer dynamics.
Infinite-dimensional control systems with outputs are considered in the
Hamiltonian formulation with generalized coordinates. An explicit scheme for
constructing a dynamic observer for this class of systems is proposed with
arbitrary gain coefficients. Sufficient conditions for the convergence of the
constructed observer are obtained on the basis of the invariance principle.
This result is applied to a flexible beam model attached to a mass-spring
system with lumped and distributed actuators. The estimation error decay is
illustrated with numerical simulations of finite-dimensional approximations of
the observer dynamics.
Authors: Alexander Zuyev, Julia Kalosha.
These applications require lower resolution, longer ranges, and
elevated viewpoints. Thus far, the dataset includes more
than 350,000 still images and over 1,300 hours of video footage of
approximately 1,000 subjects. The goal is to support the
development of algorithms capable of accurately recognizing people at ranges up
to 1,000 m and from high angles of elevation. This paper describes methods used to collect and curate the
dataset, and the dataset's characteristics at the current stage.
Face recognition technology has advanced significantly in recent years due
largely to the availability of large and increasingly complex training datasets
for use in deep learning models. These datasets, however, typically comprise
images scraped from news sites or social media platforms and, therefore, have
limited utility in more advanced security, forensics, and military
applications. These applications require lower resolution, longer ranges, and
elevated viewpoints. To meet these critical needs, we collected and curated the
first and second subsets of a large multi-modal biometric dataset designed for
use in the research and development (R&D) of biometric recognition technologies
under extremely challenging conditions. Thus far, the dataset includes more
than 350,000 still images and over 1,300 hours of video footage of
approximately 1,000 subjects. To collect this data, we used Nikon DSLR cameras,
a variety of commercial surveillance cameras, specialized long-rage R&D
cameras, and Group 1 and Group 2 UAV platforms. The goal is to support the
development of algorithms capable of accurately recognizing people at ranges up
to 1,000 m and from high angles of elevation. These advances will include
improvements to the state of the art in face recognition and will support new
research in the area of whole-body recognition using methods based on gait and
anthropometry. This paper describes methods used to collect and curate the
dataset, and the dataset's characteristics at the current stage.
Authors: David Cornett III, Joel Brogan, Nell Barber, Deniz Aykac, Seth Baird, Nick Burchfield, Carl Dukes, Andrew Duncan, Regina Ferrell, Jim Goddard, Gavin Jager, Matt Larson, Bart Murphy, Christi Johnson, Ian Shelley, Nisha Srinivas, Brandon Stockwell, Leanne Thompson, Matt Yohe, Robert Zhang, Scott Dolvin, Hector J. Santos-Villalobos, David S. Bolme.
In this paper, we focus on the theoretical analysis of diffusion-based generative modeling. Our result does not rely on any log-concavity or functional inequality assumption and has a logarithmic dependence on the smoothness. Our theoretical analysis also provides quantitative comparison between different discrete approximations and may guide the choice of discretization points in practice.
In this paper, we focus on the theoretical analysis of diffusion-based
generative modeling. Under an $L^2$-accurate score estimator, we provide
convergence guarantees with polynomial complexity for any data distribution
with second-order moment, by either employing an early stopping technique or
assuming smoothness condition on the score function of the data distribution.
Our result does not rely on any log-concavity or functional inequality
assumption and has a logarithmic dependence on the smoothness. In particular,
we show that under only a finite second moment condition, approximating the
following in KL divergence in $\epsilon$-accuracy can be done in $\tilde
O\left(\frac{d^2 \log^2 (1/\delta)}{\epsilon^2}\right)$ steps: 1) the
variance-$\delta$ Gaussian perturbation of any data distribution; 2) data
distributions with $1/\delta$-smooth score functions. Our theoretical analysis
also provides quantitative comparison between different discrete approximations
and may guide the choice of discretization points in practice.
Authors: Hongrui Chen, Holden Lee, Jianfeng Lu.
Rule-based classification models described in the language of logic directly
predict boolean values, rather than modeling a probability and translating it
into a prediction as done in statistical models. The vast majority of existing
uncertainty quantification approaches rely on models providing continuous
output not available to rule-based models. The confidence is based on how well that input
region is explored and is designed to work in any OOD scenario. We demonstrate
the usefulness of this uncertainty model by building an abstaining classifier
powered by it and observing its performance in various scenarios.
Rule-based classification models described in the language of logic directly
predict boolean values, rather than modeling a probability and translating it
into a prediction as done in statistical models. The vast majority of existing
uncertainty quantification approaches rely on models providing continuous
output not available to rule-based models. In this work, we propose an
uncertainty quantification framework in the form of a meta-model that takes any
binary classifier with binary output as a black box and estimates the
prediction accuracy of that base model at a given input along with a level of
confidence on that estimation. The confidence is based on how well that input
region is explored and is designed to work in any OOD scenario. We demonstrate
the usefulness of this uncertainty model by building an abstaining classifier
powered by it and observing its performance in various scenarios.
Authors: Yusik Kim.
In many real-world distributed settings, spurious correlations exist due to biases and data sampling issues on distributed devices or clients that can erroneously influence models. Current generalization approaches are designed for centralized training and attempt to identify features that have an invariant causal relationship with the target, thereby reducing the effect of spurious features. However, such invariant risk minimization approaches rely on apriori knowledge of training data distributions which is hard to obtain in many applications. We evaluate our approach on real-world datasets from different domains and show that FedGen results in models that achieve significantly better generalization than current federated learning approaches.
Existing federated learning models that follow the standard risk minimization
paradigm of machine learning often fail to generalize in the presence of
spurious correlations in the training data. In many real-world distributed
settings, spurious correlations exist due to biases and data sampling issues on
distributed devices or clients that can erroneously influence models. Current
generalization approaches are designed for centralized training and attempt to
identify features that have an invariant causal relationship with the target,
thereby reducing the effect of spurious features. However, such invariant risk
minimization approaches rely on apriori knowledge of training data
distributions which is hard to obtain in many applications. In this work, we
present a generalizable federated learning framework called FedGen, which
allows clients to identify and distinguish between spurious and invariant
features in a collaborative manner without prior knowledge of training
distributions. We evaluate our approach on real-world datasets from different
domains and show that FedGen results in models that achieve significantly
better generalization than current federated learning approaches.
Authors: Praveen Venkateswaran, Vatche Isahagian, Vinod Muthusamy, Nalini Venkatasubramanian.