Explore

Advertise on Engormix

Towards Automated Chicken Monitoring: Dataset and Machine Learning Methods for Visual, Noninvasive Reidentification

Published: May 8, 2025
By: Daria Kern 1,2*; Tobias Schiele 1,2; Ulrich Klauck 1,3 and Winfred Ingabire 2 / 1 Faculty Electronics & Computer Science, Aalen University, 73430 Aalen, Germany; 2 School of Computing, Engineering and Built Environment, Glasgow Caledonian University, Glasgow G4 0BA, UK; 3 Department of Computer Science, University of the Western Cape, Cape Town 7535, South Africa.
Summary

The chicken is the world’s most farmed animal. In this work, we introduce the Chicks4FreeID dataset, the first publicly available dataset focused on the reidentification of individual chickens. We begin by providing a comprehensive overview of the existing animal reidentification datasets. Next, we conduct closed-set reidentification experiments on the introduced dataset, using transformer-based feature extractors in combination with two different classifiers. We evaluate performance across domain transfer, supervised, and one-shot learning scenarios. The results demonstrate that transfer learning is particularly effective with limited data, and training from scratch is not necessarily advantageous even when sufficient data are available. Among the evaluated models, the vision transformer paired with a linear classifier achieves the highest performance, with a mean average precision of 97.0%, a top-1 accuracy of 95.1%, and a top-5 accuracy of 100.0%. Our evaluation suggests that the vision transformer architecture produces higher-quality embedding clusters than the Swin transformer architecture. All data and code are publicly shared under a CC BY 4.0 license.

Keywords: chicken; poultry; livestock; re-ID; individual identification; transformer; dataset; artificial intelligence; machine learning; computer vision

1. Introduction

1.1. Background

Chickens struggle to recognize other individuals after visible changes are applied to the comb or plumage [1]. Much like chickens are able to use visual cues to differentiate each other, artificial intelligence (AI) is capable of utilizing image or video inputs for reidentification (re-ID) purposes. Animal re-ID, the task of identifying individual animals within one (or sometimes several) species, finds applications in various fields, particularly in livestock management [2–10] and in wildlife conservation efforts, where monitoring endangered species is crucial [11–14]. Re-ID falls into one of two categories: closed set and open set. In closed-set re-ID, all individuals are known from the beginning, and those to be identified can be matched with identities of a predefined set. In open-set re-ID, the identity of the individual in question may not necessarily be part of a predefined set. It is possible to encounter completely new, undocumented individuals. Such individuals must be annotated as a new identity and, upon subsequent encounters, be accurately matched. Visual traits play a pivotal role in animal re-ID within computer vision, serving as essential markers for distinguishing individuals. Conversely, little interindividual variability poses a challenge to the re-ID task. However, the task is complex and extends beyond mere visual cues. Factors such as lighting, perspective, body changes over time, and partially obscured body parts pose additional challenges [15]. The task of re-ID is closely related to tracking, where individuals are detected and tracked across various video frames. During tracking, individuals often need to be reidentified after leaving and re-entering the field of vision. For instance, the challenge of individual chickens disappearing between frames or being occluded in later frames can be effectively addressed by chicken re-ID, which accurately recognizes the returning individual as the same chicken that left the frame.

1.2. Motivation

Automated tracking holds great potential in precision livestock farming [16]. These systems may allow for the observation of social structures and behavior, enhance welfare, and lead to more efficient animal management with minimal disruption to the livestock [17]. Chicken welfare assessment is increasingly focusing on individual animals rather than entire groups [18]. To effectively monitor, i.e., behavior, especially when dealing with many animals in a large group, it is crucial to accurately recognize and track each individual [18]. Moreover, EU regulations require individual identification to ensure traceability during disease outbreaks [19]. Traditional methods for re-ID, such as leg bands, wing tags, or backpacks with sensors, can cause significant stress to the animals and have been shown to negatively affect behavior, the immune system, and body weight [20,21]. AI-based re-ID offers a noninvasive and efficient alternative to traditional identification methods. Furthermore, identifying individual chickens is essential for behavioral studies [22], as it plays a crucial role not just in experiment oversight but also in ensuring accurate statistical analysis. The stress induced by traditional methods not only affects the animals’ well-being but also compromises the validity of research findings. AI-based identification maintains the integrity of a study while minimizing the stress on the animals. Moreover, ref. [23] recommends continuing exploring the ethological complexity of chickens in settings that are noninvasive and nonharmful, not only in commercial farming settings but also in more naturalistic settings.
Despite the significant potential of re-ID technologies, there is a notable lack of datasets that form the foundation for such advancements. Notably, as far as we are aware, no publicly available re-ID dataset for chickens currently exists, underscoring the need for development in this area. Public datasets for individual animal re-ID are scarce generally [17,24], in particular well-annotated datasets [8]. The practice of openly sharing data and code should be encouraged to enhance result comparability, but not all research data are currently made public.

1.3. Contributions

In broad terms, we performed visual closed-set re-ID of 50 individual chickens. The data used were cut-out crops of the chickens’ bodies, which excluded the background. The dataset was specifically constructed for this task, with images taken from a standing position to emphasize the plumage as the primary identifying feature.
In detail,
i We present a summary of animal re-ID studies and give a comprehensive overview of publicly available datasets.
ii We address the existing gap and introduce the first publicly available dataset for chicken re-ID: Chicks4FreeID. The dataset supports closed -and open-set re-ID, as well as semantic and instance segmentation tasks. We make this thoroughly documented dataset freely accessible to the research community and the public.
iii We evaluate a species-agnostic state-of-the-art model on our dataset through two experiments. In the first experiment, we test the model using its frozen weights, which were not trained on chicken data. In the second experiment, we fine-tune the model to adapt it specifically to our dataset.
iv We train two feature extractors from scratch in a standard supervised manner and test them on our dataset. Both models are based on transformer architectures.
v We perform additional one-shot experiments with all previously mentioned models.
vi Lastly, we make all associated code publicly available to ensure transparency and facilitate further research.

2. Related Work

2.1. Re-ID

While facial recognition is a prevalent method for reidentifying humans [25], the faces of animals can likewise serve as a means to reidentify individuals, as has previously been demonstrated for rhesus macaque [26], chimpanzee [27], cats [28], lions [29], dogs [30], giant pandas [13], and red pandas [14]. Notably, ref. [31] applied transformer-based similarity learning for chicken re-ID, using the chickens’ heads as the primary identifying feature, with a focus on a uniform group of white laying hens. Additionally, they examined the importance of different head features for visual re-ID, finding that the models prioritized characteristics such as the comb, wattles, and earlobes. However, animals frequently exhibit more distinctive visual traits beyond their faces. For example, natural markings such as stripes [32–35] and scale patterns [36] have served as prominent identifiers. Also, specific body parts can contribute to distinguishing individuals, such as the fins of dolphins [37] and sharks [38]. Similarly to how fingerprints differentiate humans, the nose prints of dogs have been utilized to uniquely identify individual dogs [39]. However, some studies aimed at identifying animals with little interindividual variability. Species exhibiting minimal or subtle visual distinctions between individuals are, for instance, (polar) bears [40,41] and elephants [42].

2.2. State of the Art

To further advance the field and aid the research community, ref. [43] released the WildlifeDatasets toolkit: an open-source toolkit for animal re-ID. It gathers publicly available animal re-ID datasets in one place in an effort to make them more easily accessible and to improve usability. Included are various tools, i.e., for data handling and processing, algorithms relevant to the task of re-ID, pretrained models, as well as evaluation methods. They address the prevailing absence of standardization across the literature and facilitate the comparability and reproducibility of results. Within their work, they also introduce a new state-of-the-art model, the MegaDescriptor, notably the first foundation model for animal re-ID. Likewise, ref. [44] presented an open-source re-ID method initially developed for sea stars, which was successfully extended to seven mammalian species without adjustments. They also reported state-of-the-art results. Moreover, ref. [45] introduced Tri-AI, a system designed for the rapid detection, identification, and tracking of individuals from a wide range of primate species. The system is capable of processing both video footage and still images.

2.3. Datasets

A review of the existing resources revealed fewer than 40 publicly available datasets for animal re-ID. This led to the conclusion that a significant number of animal species are not yet covered, including chickens. Birds in general seem to be under-represented in this domain, with only a couple of datasets available [46,47]. In fact, a noticeable focus lies on marine life [48–58]. However, cattle are the most frequently featured species [3,59–63], with much of the data collected by the same group of researchers. Despite chickens being the most widely farmed animal globally, no public re-ID dataset specifically for chickens was found. Table 1 provides a summary of the publicly accessible datasets found for animal re-ID, arranged by year. Each entry details the name of the dataset (“Dataset”), the associated publication (“Publ.”), and species focus (“Species”). “IDs” denotes the number of unique identities present within the dataset. Additionally, the total number of annotated animal instances within all images of each dataset is noted (“Annot.”). An indication (*) of whether the data were derived from video sources is given as well. For ease of access, a direct link to each dataset is provided (“Avail. at”). Although all of the datasets are publicly accessible, some are released under licenses that are relatively restrictive.
Table 1. Publicly available animal re-ID datasets, arranged by date of publication. An asterisk (*) marks data derived from video footage. “n.a.” indicates that a corresponding publication could not be found.

3. The Chicks4FreeID Dataset

3.1. Overview

The Chicks4FreeID dataset contains top-down view images of individually segmented and annotated chickens. It predominantly features female chickens (hens). For simplicity and to align with colloquial language, we refer to hens as simply “chickens” and male chickens as “roosters”. Some images also feature ducks. Each image is accompanied by a color-coded semantic segmentation mask that classifies pixel values by animal category (chicken, rooster, duck) and background, as well as binary segmentation mask(s) for the animal instance(s) depicted. Additionally, the dataset includes preprocessed cut-out crops (detailed in Section 3.4) of the respective animal instances. The Chicks4FreeID dataset can be utilized for various tasks, including re-ID, semantic segmentation, instance segmentation, and possibly anomaly detection. For each task, a distinct subset configuration was created [64]. Figure 1 gives a first overview of the dataset.
Figure 1. Dataset overview.

3.2. Collection

The data were collected manually, with no images derived from any existing dataset. Eleven private households in southern Germany were visited to photograph chickens. A total of 677 images were captured using two similar models of cameras: the “CyberShot DSC-RX100 VI” and the “CyberShot DSC-RX100 I” from Sony. The resolution of the images stands at 3648 × 5472 pixels. Each image includes at least one chicken, ensuring no images without chickens are part of the dataset. The animals were allowed to move freely without constraints to prevent distress. As a result, other individuals frequently entered the frame. The chickens were photographed from a standing position, aiming to capture the plumage from a top-down perspective. However, since the animals were free to move, their distance from the camera varied, which affected the view of their plumage. Thus, the resulting images do not show a clean top-town view. Most individuals were photographed outdoors under natural light, except for four individuals, which were indoors in a coop during winter. Data collection took approximately one year. However, all photos of a given individual were taken on the same day.

3.3. Annotation

All annotations were manually created by a human annotator using Labelbox [93] under a free educational license.
Instance Masks By drawing an almost pixel-perfect polygon outline, a binary instance mask was created for each animal in an image. This process resulted in a total of 1270 instance masks. In each mask, the instance is encoded in white, whereas the black area is considered background. Each instance includes the comb, head, beak, and plumage. In contrast, feet and scattered feathers are classified as part of the background, along with any other visible objects. Feet were intentionally excluded to prevent identification through leg rings. We created instance masks instead of bounding boxes to aid the creation of precise cut-out crops for re-ID. Bounding boxes often include other individuals, which could confuse the model, as well as background elements, which could introduce unintended clues. By using instance masks, we ensure a focus on the animal’s features, avoiding reliance on background information.
Identities Each instance was assigned an identity. There were 54 distinct identities. Prior to the photography sessions, the visible features of each animal were meticulously studied and recorded in a notebook. This was conducted to ensure the correct identification of each individual during annotation. Ground truth annotation was therefore performed using expert knowledge, without the use of any algorithms. In cases where the human annotator could not assign an identity, the instance in question was labeled as identity “Unknown”. It is essential to clarify that the label “Unknown” does not imply the presence of a new, undocumented individual. Instead, it represents an unidentified individual from the closed set, more precisely, from the annotated coop. A key strength of the Chicks4FreeID dataset is its expert ground truth labeling for all instances, ensuring reliable verification of results even in open-set Re-ID tasks. This sets it apart from datasets that include completely unannotated instances or rely on ground truth labels generated through clustering or AI methods, which may introduce labeling errors.
Animal Category Each instance was assigned to one of 3 animal categories. These are “chicken”, “rooster”, and “duck”. Roosters, and particularly ducks, are exceptions in this dataset, which is primarily composed of female chickens (hens). For simplicity and to align with colloquial language, instances of female chickens were categorized as “chicken” while male chickens were categorized as “rooster”.
Visibility Rating It was common for animals to be partially obscured by other animals or objects in the images. Additionally, they were often not fully contained within the image frame. Acknowledging the resulting varying visibility of the instances, each was assigned one of 3 visibility ratings: “bad”, “good”, or “best” (for examples, see Figure A1; for full overview, see Tables A1 and A2). The “best” rating includes instances that fully display the animal from the desired top-down perspective and those where only an insignificant part is missing, such as the very tip of the tail feathers. Instances that include only small parts of the animal and with an undesired perspective fall under the “bad” rating. All remaining instances that do not qualify as “bad” or “best” are rated as “good”.
Coops The location (one of 11 coops) of each photograph was documented during the capture process. As a result, each image was annotated with the coop it belonged to. It is therefore straightforward to match the animals in the images to their respective coops.
Semantic Segmentation Masks Furthermore, 677 color-coded semantic segmentation masks were created, 1 for each image. They include four possible classes: “background”, “chicken”, “duck” and “rooster”. Every mask includes “background” and “chicken”, while “duck” and “rooster” are included if present. While the semantic segmentation masks are not directly involved in the re-ID process, they were added to the dataset due to the availability of existing annotations, expanding the dataset’s utility to support additional semantic segmentation tasks.

3.4. Preprocessing

The Chicks4FreeID dataset contains preprocessed data for the task of re-ID. These are cut-out crops of isolated animal instances with a solid black background. By removing visual cues like feet with rings or environmental features, it is ensured that the re-ID relies solely on the animals’ physical characteristics. The following steps describe the preprocessing procedure to obtain the cut-out crops for the re-ID task. For all individuals captured in an image, a bounding box is created based on the instance masks. In the first step, both the image and the mask are cropped (to the area of interest contained in the bounding box) to focus solely on the individual (see Figure 2: Step 1). The cropped mask is then used to remove the background from the cropped image (Step 2). Finally, the resulting image is adjusted to a square shape for ease of use and consistency (Step 3). The resulting resolutions remain as is, with no resizing taking place.
Figure 2. Data preprocessing pipeline for subsequent re-ID.

3.5. Dataset Statistics

The dataset contains 677 images featuring 54 individual animals. The 677 images were captured across 11 different coops, and the 54 individual animals were divided into three categories: “chicken”, “rooster”, and “duck”. There are 50 “chicken”, 2 “rooster”, and 2 “duck” identities. For simplicity and to align with common usage, hens were categorized as “chicken” while the two roosters were given their own category. Each image is paired with a semantic segmentation mask and an instance segmentation mask for each animal depicted. In total, there are 1270 instance segmentation masks. From the instance masks and the corresponding images, 1270 cut-out crops of the individual animals were generated, specifically for use in the re-ID process. There are 1215 “chicken” instances, 15 “rooster” instances, and 40 “duck” instances. The instances also received a visibility rating: 798 “best”, 201 “good”, and 271 “bad” instances. Among the 271 “bad” instances, 69 are labeled “Unknown” as they could not be certainly identified by the annotator. No “Unknown” instances are present among those rated as “best” or “good”. Therefore, the ground truth identity for all instances with these ratings is known. Figure 3 illustrates the number of instances for each individual, along with their visibility ratings. The figure starts with the individual having the most instances rated as “best” and is arranged in descending order. The chicken identity with the most “best” ratings is “Mirmir”, with 27 instances, while “Isolde” has the fewest, with 4 “best” rated instances. The dataset comprises chicken individuals with both uniform and nonuniform plumage. The diversity in appearance possibly aids individual identification, with differences in colors and patterns serving as markers. As a result, this renders the dataset easier to solve than one containing only uniform-plumage individuals. Nevertheless, some individuals in the dataset exhibit uniform plumage, including four that are solid white, four that are solid black, four in shades of gray, and six in shades of orange. Examples of these plumage colors are illustrated in Figure A2. Detailed information about the individuals and the corresponding annotations is provided in Table A3.
Figure 3. Visibility distributions for all instances of each individual. Ducks and roosters are marked with an asterisk (*).

4. Materials and Methods

4.1. Hardware

The training and evaluation were conducted on 64 GB shared memory Apple M3 Max Chips (2023) running PyTorch 2.3.0 with MPS acceleration.

4.2. Data

For the experiments, we employed the subset “chicken-re-id-all-visibility”, which features 50 different individuals. It is hosted on Hugging Face [64], a widely used platform for machine learning datasets and models. The dataset encompasses 1146 “chicken” instances of all visibility ratings (793 “best”, 181 “good”, and 172 “bad”). It does not contain any “Unknown” instances nor instances of the animal category “rooster” or “duck”. As a result, it comprises 1146 pairs of preprocessed “chicken” cut-out crops and assigned identities. The subset was split into 916 train pairs and 230 test pairs (stratified). All 50 identities were included in the training set; this ensured that the testing set did not introduce any new identities. For a fair evaluation on all identities, the train/test split was stratified, i.e., each identity had the same fixed percentage of its cut-out crops allocated to the test set. Hence, the test set also included all of the 50 identities. As a consequence of the stratified split, identities with a higher total number of crops contributed more to the test set compared to identities with fewer crops, ensuring proportional representation across all identities. After accessing the described “chicken-re-id-all-visibility”subset, we additionally introduced a validation split (10% of the training set, stratified random split). Consequently, the dataset was finally split into 824 pairs for training (prior to applying any augmentation), 92 pairs for validation, and 230 pairs for testing.

4.3. Augmentation

Augmentation was applied dynamically during the training process to the training data: random rotation (360 degrees), random flip (horizontal and vertical), and RandAugment [94]. No data augmentation was applied to the test or validation set. To avoid data leakage, it is important to apply data augmentation only after a train-test split was established. This ensures that augmented versions of the same original image do not appear in both sets.

4.4. Feature Extractors

Vision Transformer We employed the ViT-B/16 [95] architecture, as implemented in [96]. The hyperparameters (ADAM + CosineWarmup) were inspired by the optimizer used in the lightly benchmarks [97] for their vision transformer backbones.
Swin Transformer We also utilized the swin_large_patch4_window12_384 architecture [98] as implemented in [99]. The hyperparameters (Stochastic Gradient Descent + Cosine Annealing, ArcFace [100] loss function) mirrored those used to build the MegaDescriptor, which also employs the same Swin transformer architecture. The Swin transformer itself is based on the vision transformer architecture [95]. The difference between the Swin Transformer and the vision transformer lies in how they handle image data: the Swin transformer uses a hierarchical structure with shifted windows to capture local and global features, while the vision transformer treats images as sequences of patches, relying on self-attention mechanisms throughout.
MegaDescriptor The employed MegaDescriptor-L-384 [43] (CC BY-NC 4.0 license [101]) is a state-of-the-art feature extractor for animal re-ID from the WildlifeDatasets toolkit (MIT license). It is based on the Swin transformer architecture [98] and was pretrained on diverse datasets featuring various animal species. A notable hyperparameter choice made by the MegaDescriptor-L384 authors is the ArcFace [100] loss function, which aims to aid in building meaningful embeddings. We selected the frozen MegaDescriptor-L-384 model over DINOv2 [102] and CLIP [103] due to its better performance on unseen animal domains, as reported by the authors. Their evaluation included cattle as an example of an unseen domain [43].

4.5. Classifiers

k-NN k-NN stores all the training data and calculated distances at the time of prediction. It works by comparing a new chicken’s embedding (feature vector) to all known embeddings. It then finds the k-closest embeddings in the feature space and assigns the most common identity among those neighbors. Our settings were inspired by [104]. We used a high value of k = 200 to provide a broader context, allowing us to consider not just the closest neighbor but the overall composition of the embedding space. For a correct prediction, other individuals must be sufficiently distant from the cluster of the “correct” individual. This k-NN approach prioritizes clustering quality over absolute discriminative power.
Linear Classifier To evaluate absolute discriminative power, we relied on a linear classifier. The employed linear classifier is a simple linear layer. It was trained for 90 epochs using the extracted embeddings (feature vectors) of the training data. Thereby, it learned to associate each feature vector with the correct identity by adjusting its weights, ultimately mapping each input to1one of the 50 identity classes. It was validated on the embeddings of the validation set.

4.6. Evaluation Metrics

We provide three of the most common metrics for closed set animal re-ID. These are mAP (mean average precision), top-1 accuracy (ratio of correct predictions versus total predictions), and top-5 accuracy (accuracy of the correct class being within the top 5 predictions) as implemented in TorchMetrics [105]. It is important to note that the mAP focuses more on the quality of the probability estimates, serving as a ranking metric that assesses “how well the model assigns high probabilities to the correct chicken.” This is similar to the top-5 accuracy, where the model’s prediction is considered correct if the correct answer is among the top five guesses. However, even though top-5 accuracy can provide some confidence in the model’s predictions, it does not fully capture the model’s actual performance in terms of the number of errors made. Therefore, the top-1 accuracy must be considered the most.

5. Experiments

The approach for closed-set re-ID involved two steps. First, a feature extractor generated embeddings for the cut-out crops. Second, the resulting feature vectors (embeddings) were then passed to a classifier to ultimately assign the identities. Feature extractor and classifier were trained separately. Figure 4 illustrates the training and evaluation process. We evaluated four feature extraction models: MegaDescriptor (both frozen and fine-tuned), vision transformer, and Swin transformer. All feature extractors were fed with images at an input resolution of 384 × 384 pixels (the square-shaped cut-out crops were resized). Additionally, we tested with a variation of two classifiers: k-NN and a linear classifier. In our experiments, we explored different training strategies: standard supervised learning (see Section 5.2) and one-shot learning (see Section 5.3). We employed both standard supervised learning and one-shot learning approaches to evaluate performance across different data availability scenarios. The one-shot setting simulated situations with very limited data while also providing insights into generalization capabilities by reducing the risk of overfitting. We also specifically compared the performance of the pretrained MegaDescriptor with its fine-tuned counterpart (see Section 5.1). Each experiment, as a whole, was repeated three times to ensure consistency.
Figure 4. Illustration of the training and evaluation process for the feature extractor and classifier, showcasing the linear classifier as an example in this workflow.

5.1. Domain Transfer Experiment

This experiment had two main objectives. First, it aimed to evaluate the performance of the frozen pretrained MegaDescriptor on our Chicks4FreeID dataset, which represents a new, unseen domain. The MegaDescriptor was originally trained on diverse datasets featuring various animal species. However, chickens were not part of the original training data. We evaluated the MegaDescriptor as a feature extractor with its weights frozen, preserving its pretrained state. Second, the experiment sought to enhance performance by fine-tuning the MegaDescriptor on the Chicks4FreeID dataset. This optimally allows the model to adapt by leveraging the features it previously learned from other species and refining them through additional training. For this, we unfroze all the model’s layers and trained it for 200 additional epochs using the training set, as detailed in Section 4.2. For testing and validation, we utilized the test and validation sets, also described in Section 4.2.

5.2. Standard Supervised Learning Experiment

This experiment aimed to train a chicken feature extractor from scratch using the Chicks4FreeID dataset. We compared two transformer architectures; the Swin and the vision transformers.

5.3. One-Shot Experiment

This experiment aimed to evaluate all previously discussed models in a one-shot learning setting, where the training set was reduced to only one sample per individual. With 50 individuals in the dataset, the training set initially consisted of 50 samples, each representing one individual. All remaining samples of the previously established training set (see Section 4.2) were unused. These 50 samples were then augmented dynamically during training using the methods described in Section 4.3. The validation and test sets remained unchanged, as outlined in Section 4.2. All feature extractors, except the non-fine-tuned frozen MegaDescriptor, were trained on this one-shot training set. The classifiers were also trained/fitted solely on this reduced set. For the non-fine-tuned frozen MegaDescriptor, the only difference in the one-shot setting was therefore the classifier. Among the conducted experiments, this scenario most closely resembled an open-set re-ID setting, where a new, previously undocumented identity appears and initially little data are available. Consequently, the expected performance in an open-set scenario is likely to be similar the results observed in the one-shot setting, particularly for the non-fine-tuned frozen MegaDescriptor. However, the classifier would need to be retrained each time a new individual appears. To enable open-set re-ID and to identify unknown individuals without retraining, the classifier head could be replaced with a clustering algorithm, such as DBSCAN [106,107], or hierarchical clustering [108].

6. Results and Discussion

Tables 2–4 show the results on the test data for the domain transfer, standard supervised learning, and one-shot experiments, respectively. Figure 5 additionally illustrates the top-1 accuracy as bar charts. For runtime results and discussion, see Figure A3 in the Appendix A.
Table 2. Domain transfer experiment results. The highest scores for each metric are in blue. Rows corresponding to experiments employing a linear classifier are highlighted in gray.
Table 3. Results of standard supervised learning experiment. Highest scores for each metric are in blue. Rows corresponding to experiments employing a linear classifier are highlighted in gray.
Table 4. One-shot learning experiment results. Highest scores for each metric are in blue. Rows corresponding to experiments employing a linear classifier are highlighted in gray.
Figure 5. Top-1 accuracy visualized in bar charts. The left bar chart combines the results in Tables 2 and 3. The right bar chart illustrates the one-shot experiments in Table 4.
MegaDescriptor Shows Improved Performance with Fine-Tuning The fine-tuned MegaDescriptor model achieved a mAP of 96.0% and a top-1 accuracy of 91.6% when paired with a linear classifier (see Table 2). It outperformed the frozen MegaDescriptor in the domain transfer experiment. This aligns with expectations that a model performs not as well on an unseen domain. The fine-tuned MegaDescriptor also outperformed the Swin transformer, which we specifically trained on our data from scratch (see Table 3). This suggests the pretrained features are useful for identifying chickens, since fine-tuning seems preferable to training the same model from scratch.
Frozen MegaDescriptor is a Powerful Feature Extractor The frozen MegaDescriptor, when combined with a linear classifier, still performed well, with a mAP of 92.0% and a top-1 accuracy of 81.8% (see Table 2). Its performance was not far behind that of models specifically trained on our data (see Table 3). This is particularly impressive given that the frozen model had no exposure to chicken data. It was trained on a large dataset of many different animal species, which also included bird images. Since birds share common features like feathers, this overlap may have further contributed to the model’s strong performance, in addition to the large training dataset.
Trained-from-Scratch Vision Transformer Yields the Best Results The vision transformer trained from scratch, combined with a linear classifier, achieved the overall best results with a mAP of 97.0%, a top-1 accuracy of 95.1%, and a perfect top-5 accuracy of 100.0% (see Table 3).
Fine-Tuned MegaDescriptor Leads in One-shot Setting As expected, the performance dropped for all models in the one-shot setting (see Table 4). When trained with only a single image per individual, the fine-tuned MegaDescriptor achieved the best performance, with a 64.5% mAP, a 56.2% top-1 accuracy, and an 81.8% top-5 accuracy, closely followed by the vision transformer, which achieved a 61.9% mAP, a 52.3% top-1 accuracy, and a 80.8% top-5 accuracy. Both models were paired with a linear classifier for these evaluations. The fine-tuned MegaDescriptor likely outperformed the vision transformer in this scenario because it had already learned embeddings through pretraining before being fine-tuned in a one-shot manner. In contrast, the vision transformer had no prior learning, starting from scratch. This pretraining advantage gave the MegaDescriptor a head start in understanding features, likely contributing to its better performance. This underscores that in scenarios with very limited data, like the one-shot example, transfer learning (fine-tuning pretrained models) is more effective than training from scratch. Furthermore, we think that the improved discrimination capabilities make the fine-tuned MegaDescriptor more suitable in an open-set re-ID setting than the other models.
Evaluation Focus Varies Between Linear Classifiers and k-NN In all experiments, feature extractors consistently performed worse when paired with k-NN compared to when they were paired with a linear classifier. This was mainly due to the differences in the evaluation focus of the employed classifiers. The employed k-NN approach is more challenging by design as it aims to evaluate the clustering quality in embedding space, instead of absolute discriminative power. Additionally, the k-NN is a parameterless model, while the linear classifier has approximately hidden_dim × 50 parameters. k-NN only captures distances in the feature vector space, whereas a linear classifier can learn robust decision boundaries. Moreover, with a linear classifier, we can track training progress with a validation split, leading to better generalization compared to the fixed boundaries of k-NN.
Higher-quality Embedding Clusters with Vision Transformer Training the Swin transformer architecture, which also served as the basis for the MegaDescriptor, appeared to result in lower clustering quality of the embeddings compared to the vision transformer architecture, meaning the overall composition of the embedding space appeared to be better when employing a vision transformer. This difference is evident in Tables 2 and 3 when paired with k-NN. This could be due to the simpler architecture and loss function, which could lead to reaching better performance faster.
Higher Accuracy with Vision Transformer Furthermore, the vision transformer consistently outperformed the Swin transformer when paired with a linear classifier. This could be because it is optimized to predict class probabilities directly, while the Swin transformer focuses on learning the relative distances between features in the feature space. This difference in optimization goals allows the vision transformer to better capture the overall structure of the data, leading to more accurate classification when paired with the same classifier.

7. Conclusions

7.1. Findings

We introduced the first open-source dataset for chicken re-ID to address the need for well-annotated, publicly available datasets for animal re-ID. Utilizing this dataset, we conducted a series of closed-set re-ID experiments. We combined various transformer-based feature extractors (MegaDescriptor, vision transformer, Swin transformer) with two classifiers (k-NN and linear classifier) and evaluated their performance across different settings (domain transfer, training from scratch, and one-shot learning).
We found that pretraining and domain transfer are critical when dealing with limited data per instance. The MegaDescriptor model proved well-suited as a pretrained model in such scenarios. Pretraining on a different domain (e.g., animals other than chickens) enables a model to learn generalizable features that can later be fine-tuned to the specific target domain (e.g., chickens). Interestingly, if sufficient data are available, training a feature extractor from scratch is not necessarily preferable, as the fine-tuned MegaDescriptor outperformed the Swin transformer, which was trained from scratch.
Furthermore, our evaluation suggests that the vision transformer architecture produces higher-quality embedding clusters than the Swin transformer architecture.
Overall, transformer-based architectures as feature extractors achieved high accuracy and mAP on our dataset. The best results on our dataset, however, were achieved with the vision transformer trained from scratch combined with a linear classifier (mAP 97.0%, top-1 95.1%, top-5 100.0%).

7.2. Limitations and Future Work

We have not yet explored which features the models prioritize during the identification process, such as patterns, colors, shapes, specific areas of the plumage, or certain body parts. This presents an interesting direction for future research. The classifiers used for the experiments were straightforward and effective for the purposes of this study. While more advanced classifiers could potentially yield even better results, the main focus of this work was on evaluating the feature extractors. Another potential direction for future work is adapting the Chicks4FreeID dataset for open-set Re-ID experiments. Researchers could, for example, exclude certain individuals from the training set and introduce them later during evaluation. The dataset subset used for the experiments contained 1215 instance annotations of 50 individual chickens, offering a compact yet valuable resource. The versioning system of the dataset facilitates potential expansions and continuous improvements, ensuring its ongoing relevance and applicability for future research. For chicken breeds with minimal interindividual variability, such as those with uniform plumage, increasing both the number of individuals and instances per individual could further improve the re-ID process. This expansion would be particularly relevant for applications in industrial farming, where thousands of chickens of a single breed are typically kept. Furthermore, photographing individual chickens over an extended period proved challenging, as free-range chickens are often preyed upon by wild animals such as raccoons, foxes, or rats. Therefore, the decision was made to capture all images of a given individual in a single day. Unfortunately, this approach did not capture changes in appearance over time, such as molting. Future work could focus on increasing variability by photographing chickens under different lighting conditions, such as from dusk till dawn. It could also explore the use of different imaging techniques, like multispectral imaging. Additional improvements could include capturing different angles (front, back, side), documenting various stages of life, and recording changes over time. While challenging, these additions could provide a more comprehensive dataset. However, a dataset gains its true value when it is made publicly available. Therefore, we encourage the release of future datasets under open, unrestricted licenses.

Appendix A. Annotations

Figure A1. Examples of visibility rating “best”, “good”, and “bad”.
Table A1. Full overview of all “chicken” annotations in the Chicks4FreeID dataset.
Table A2. Full overview of all “rooster” and “duck” annotations in the Chicks4FreeID dataset.
Table A2. Full overview of all “rooster” and “duck” annotations in the Chicks4FreeID dataset.

Appendix B. Plumage

Figure A2. Uniform plumage examples from left to right: solid white, solid black, shades of gray, shades of orange.
Figure A2. Uniform plumage examples from left to right: solid white, solid black, shades of gray, shades of orange.
Table A3. Overview of uniform plumage “chicken” individuals in the Chicks4FreeID dataset. Grand total of annotations of mixed plumage individuals is included for completeness.
Table A3. Overview of uniform plumage “chicken” individuals in the Chicks4FreeID dataset. Grand total of annotations of mixed plumage individuals is included for completeness.

Appendix C. Runtime

For efficiency and performance comparison, we report runtime speeds for each run in Figure A3.
Figure A3. Mean runtime results by experiment type and model (log scale).
Embedding training runtime and classifier training runtime are reported separately. Since each experiment was repeated three times, we report the mean values and error bars for each distinct experiment group. k-NN was the fastest classifier, taking around 1 min per run. This was consistent across both one-shot and full experiments, indicating that runtime speed was primarily dominated by overhead. Overall, excluding k-NN, training one-shot models was significantly faster than training with full datasets, by one order of magnitude, as seen in the logarithmic scale. The figure also shows that training only the classifiers (k-NN and linear) was much more efficient than training or fine-tuning the full pipeline. Training the comparatively large feature extractors took considerably longer. Another noticeable gap exists between the vision transformer and MegaDescriptor architectures (see Embedding Training bars for MegaDescriptor (fine-tuned) and Swin transformer versus vision transformer). This difference arises because the vision transformer avoids calculating shifting windows and uses a simpler loss function that requires only one image instead of two.
    
This article was originally published in Animals 2025, 15, 1. https://doi.org/10.3390/ ani15010001. This is an Open Access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).

1. Guhl, A.M.; Ortman, L.L. Visual Patterns in the Recognition of Individuals among Chickens. Condor 1953, 55, 287–298. [CrossRef]

2. Andrew, W.; Hannuna, S.; Campbell, N.; Burghardt, T. Automatic individual holstein friesian cattle identification via selective local coat pattern matching in RGB-D imagery. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 484–488. [CrossRef]

3. Li, S.; Fu, L.; Sun, Y.; Mu, Y.; Chen, L.; Li, J.; Gong, H. Cow Dataset. 2021. Available online: https://doi.org/10.6084/m9.figshare. 16879780 (accessed on 3 November 2024).

4. Andrew, W.; Greatwood, C.; Burghardt, T. Visual Localisation and Individual Identification of Holstein Friesian Cattle via Deep Learning. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 2850–2859. [CrossRef]

5. Gao, J.; Burghardt, T.; Andrew, W.; Dowsey, A.W.; Campbell, N.W. Towards Self-Supervision for Video Identification of Individual Holstein-Friesian Cattle: The Cows2021 Dataset. arXiv 2021, arXiv:2105.01938 .

6. Andrew, W.; Gao, J.; Mullan, S.; Campbell, N.; Dowsey, A.W.; Burghardt, T. Visual identification of individual Holstein-Friesian cattle via deep metric learning. Comput. Electron. Agric. 2021, 185, 106133. [CrossRef]

7. Zhang, T.; Zhao, Q.; Da, C.; Zhou, L.; Li, L.; Jiancuo, S. YakReID-103: A Benchmark for Yak reidentification. In Proceedings of the 2021 IEEE International Joint Conference on Biometrics (IJCB), Shenzhen, China, 4–7 August 2021; pp. 1–8. [CrossRef]

8. Chan, J.; Carrión, H.; Mégret, R.; Rivera, J.L.A.; Giray, T. Honeybee reidentification in Video: New Datasets and Impact of Self-supervision. In Proceedings of the 17th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2022)—Volume 5: VISAPP, INSTICC, Avenida de S. Francisco Xavier, Lote 7 Cv. C, Online, 6–8 February 2022; pp. 517–525. [CrossRef]

9. Tausch, F.; Stock, S.; Fricke, J.; Klein, O. Bumblebee reidentification Dataset. In Proceedings of the 2020 IEEE Winter Applications of Computer Vision Workshops (WACVW), Snowmass Village, CO, USA, 1–5 March 2020; pp. 35–37. [CrossRef]

10. Borlinghaus, P.; Tausch, F.; Rettenberger, L. A Purely Visual Re-ID Approach for Bumblebees (Bombus terrestris). Smart Agric. Technol. 2023, 3, 100135. [CrossRef]

11. Kulits, P.; Wall, J.; Bedetti, A.; Henley, M.; Beery, S. ElephantBook: A Semi-Automated Human-in-the-Loop System for Elephant reidentification. In Proceedings of the 4th ACM SIGCAS Conference on Computing and Sustainable Societies, New York, NY, USA, 28 June–2 July 2021; pp. 88–98. [CrossRef]

12. Moskvyak, O.; Maire, F.; Dayoub, F.; Armstrong, A.O.; Baktashmotlagh, M. Robust reidentification of Manta Rays from Natural Markings by Learning Pose Invariant Embeddings. In Proceedings of the 2021 Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, Australia, 29 November–1 December 2021; pp. 1–8. [CrossRef]

13. Wang, L.; Ding, R.; Zhai, Y.; Zhang, Q.; Tang, W.; Zheng, N.; Hua, G. Giant Panda Identification. IEEE Trans. Image Process. 2021, 30, 2837–2849. [CrossRef]

14. He, Q.; Zhao, Q.; Liu, N.; Chen, P.; Zhang, Z.; Hou, R. Distinguishing Individual Red Pandas from Their Faces. In Pattern Recognition and Computer Vision; Lin, Z., Wang, L., Yang, J., Shi, G., Tan, T., Zheng, N., Chen, X., Zhang, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 714–724.

15. Vidal, M.; Wolf, N.; Rosenberg, B.; Harris, B.P.; Mathis, A. Perspectives on Individual Animal Identification from Biology and Computer Vision. Integr. Comp. Biol. 2021, 61, 900–916. [CrossRef]

16. Neethirajan, S. ChickTrack—A quantitative tracking tool for measuring chicken activity. Measurement 2022, 191, 110819. [CrossRef]

17. T. Psota, E.; Schmidt, T.; Mote, B.; C. Pérez, L. Long-Term Tracking of Group-Housed Livestock Using Keypoint Detection and MAP Estimation for Individual Animal Identification. Sensors 2020, 20, 3670. [CrossRef]

18. Li, N.; Ren, Z.; Li, D.; Zeng, L. Review: Automated techniques for monitoring the behaviour and welfare of broilers and laying hens: Towards the goal of precision livestock farming. Animal 2020, 14, 617–625. [CrossRef]

19. Food Traceability. European Commission—B-1049 Brussels. 2007. Available online: https://food.ec.europa.eu/system/files/20 16-10/gfl_req_factsheet_traceability_2007_en.pdf (accessed on 13 August 2024).

20. Dennis, R.L.; Fahey, A.G.; Cheng, H.W. Different Effects of Individual Identification Systems on Chicken Well-Being1. Poult. Sci. 2008, 87, 1052–1057. [CrossRef]

21. Anderson, G.; Johnson, A.; Arguelles-Ramos, M.; Ali, A. Impact of Body-worn Sensors on Broiler Chicken Behavior and Agonistic Interactions. J. Appl. Anim. Welf. Sci. 2023, 1–10. [CrossRef] [PubMed]

22. Stadig, L.M.; Rodenburg, T.B.; Ampe, B.; Reubens, B.; Tuyttens, F.A. An automated positioning system for monitoring chickens’ location: Effects of wearing a backpack on behaviour, leg health and production. Appl. Anim. Behav. Sci. 2018, 198, 83–88. [CrossRef]

23. Marino, L. Thinking chickens: A review of cognition, emotion, and behavior in the domestic chicken. Anim. Cogn. 2017, 20, 127–147. [CrossRef] [PubMed]

24. Lu, W.; Zhao, Y.; Wang, J.; Zheng, Z.; Feng, L.; Tang, J. MammalClub: An Annotated Wild Mammal Dataset for Species Recognition, Individual Identification, and Behavior Recognition. Electronics 2023, 12, 4506. [CrossRef]

25. Ye, M.; Shen, J.; Lin, G.; Xiang, T.; Shao, L.; Hoi, S.C.H. Deep Learning for Person reidentification: A Survey and Outlook. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 2872–2893. [CrossRef]

26. Witham, C.L. Automated face recognition of rhesus macaques. J. Neurosci. Methods 2018, 300, 157–165. [CrossRef]

27. Freytag, A.; Rodner, E.; Simon, M.; Loos, A.; Kühl, H.S.; Denzler, J. Chimpanzee Faces in the Wild: Log-Euclidean CNNs for Predicting Identities and Attributes of Primates. In Pattern Recognition; Rosenhahn, B., Andres, B., Eds.; Springer: Cham, Switzerland, 2016; pp. 51–63.

28. Lin, T.Y.; Kuo, Y.F. Cat Face Recognition Using Deep Learning; American Society of Agricultural and Biological Engineers (ASABE) Annual International Meeting: St. Joseph, MI, USA, 2018. [CrossRef]

29. Dlamini, N.; Zyl, T.L.v. Automated Identification of Individuals in Wildlife Population Using Siamese Neural Networks. In Proceedings of the 2020 7th International Conference on Soft Computing and Machine Intelligence (ISCMI), Stockholm, Sweden, 14–15 November 2020; pp. 224–228. [CrossRef]

30. Mougeot, G.; Li, D.; Jia, S. A Deep Learning Approach for Dog Face Verification and Recognition. In Proceedings of the PRICAI 2019: Trends in Artificial Intelligence, Cuvu, Fiji, 26–30 August 2019; Nayak, A.C.; Sharma, A., Eds.; Springer International Publishing: Cham, Swtizerland, 2019; pp. 418–430. [CrossRef]

31. Lamping, C.; Kootstra, G.; Derks, M. Transformer-Based Similarity Learning for Re-Identification of Chickens. Available online: https://doi.org/10.2139/ssrn.4886408 (accessed on 22 October 2024).

32. Li, S.; Li, J.; Tang, H.; Qian, R.; Lin, W. ATRW: A Benchmark for Amur Tiger Re-identification in the Wild. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2590–2598. [CrossRef]

33. Parham, J.; Crall, J.; Stewart, C.; Berger-Wolf, T.; Rubenstein, D. Animal population censusing at scale with citizen science and photographic identification. Aaai Spring Symp. Tech. Rep. 2017, SS-17-01–SS-17-08, 37–44.

34. Haurum, J.B.; Karpova, A.; Pedersen, M.; Bengtson, S.H.; Moeslund, T.B. reidentification of Zebrafish using Metric Learning. In Proceedings of the 2020 IEEE Winter Applications of Computer Vision Workshops (WACVW), Snowmass Village, CO, USA, 1–5 March 2020; pp. 1–11. [CrossRef]

35. Lahiri, M.; Tantipathananandh, C.; Warungu, R.; Rubenstein, D.I.; Berger-Wolf, T.Y. Biometric animal databases from field photographs: Identification of individual zebra in the wild. In Proceedings of the 1st ACM International Conference on Multimedia Retrieval, Vancouver, BC, Canada, 30–31 October 2008. [CrossRef]

36. Adam, L.; Cermák, V.; Papafitsoros, K.; Picek, L. SeaTurtleID2022: A long-span dataset for reliable sea turtle reidentification. In ˇ Proceedings of the 2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 7131–7141. [CrossRef]

37. Bouma, S.; Pawley, M.D.; Hupman, K.; Gilman, A. Individual Common Dolphin Identification via Metric Embedding Learning. In Proceedings of the 2018 International Conference on Image and Vision Computing New Zealand (IVCNZ), Auckland, New Zealand, 19–21 November 2018; pp. 1–6. [CrossRef]

38. Hughes, B.; Burghardt, T. Automated Visual Fin Identification of Individual Great White Sharks. Int. J. Comput. Vision 2017, 122, 542–557. [CrossRef]

39. Bae, H.B.; Pak, D.; Lee, S. Dog Nose-Print Identification Using Deep Neural Networks. IEEE Access 2021, 9, 49141–49153. [CrossRef]

40. Zuerl, M.; Dirauf, R.; Koeferl, F.; Steinlein, N.; Sueskind, J.; Zanca, D.; Brehm, I.; Fersen, L.v.; Eskofier, B. PolarBearVidID: A Video-Based reidentification Benchmark Dataset for Polar Bears. Animals 2023, 13, 801. [CrossRef]

41. Clapham, M.; Miller, E.; Nguyen, M.; Darimont, C.T. Automated facial recognition for wildlife that lack unique markings: A deep learning approach for brown bears. Ecol. Evol. 2020, 10, 12883–12892. [CrossRef] [PubMed]

42. Körschens, M.; Denzler, J. ELPephants: A Fine-Grained Dataset for Elephant reidentification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 263–270. [CrossRef]

43. Cermák, V.; Picek, L.; Adam, L.; Papafitsoros, K. WildlifeDatasets: An Open-Source Toolkit for Animal reidentification. In ˇ Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 5953–5963.

44. Wahltinez, O.; Wahltinez, S.J. An open-source general purpose machine learning framework for individual animal reidentification using few-shot learning. Methods Ecol. Evol. 2024, 15, 373–387. [CrossRef]

45. Guo, S.; Xu, P.; Miao, Q.; Shao, G.; Chapman, C.A.; Chen, X.; He, G.; Fang, D.; Zhang, H.; Sun, Y.; et al. Automatic Identification of Individual Primates with Deep Learning Techniques. iScience 2020, 23, 32. [CrossRef] [PubMed]

46. Ferreira, A.C.; Silva, L.R.; Renna, F.; Brandl, H.B.; Renoult, J.P.; Farine, D.R.; Covas, R.; Doutrelant, C. Bird individualID. 2020. Available online: https://github.com/AndreCFerreira/Bird_individualID (accessed on 3 November 2024).

47. Kuncheva, L.I.; Williams, F.; Hennessey, S.L.; Rodríguez, J.J. Animal-Identification-from-Video. 2022. Available online: https: //github.com/LucyKuncheva/Animal-Identification-from-Video (accessed on 3 November 2024).

48. Adam, L.; Cermák, V.; Papafitsoros, K.; Picek, L. SeaTurtleID. 2022. Available online: ˇ https://www.kaggle.com/datasets/ wildlifedatasets/seaturtleid2022 (accessed on 3 November 2024).

49. Wahltinez, O. Sea Star Re-ID. 2023. Available online: https://lila.science/sea-star-re-id-2023/ (accessed on 3 November 2024).

50. Me, W. Beluga ID. info@wildme.org. 2022. Available online: https://lila.science/datasets/beluga-id-2022/ (accessed on 3 November 2024).

51. Cheeseman, T.; Southerland, K.; Reade, W.; Howard, A. Happywhale—Whale and Dolphin Identification. 2022. Available online: https://kaggle.com/competitions/happy-whale-and-dolphin (accessed on 3 November 2024).

52. Nepovinnykh, E. SealID. Lappeenranta University of Technology, School of Engineering Science Yhteiset. 2022. Available online:. (accessed on 3 November 2024). [CrossRef]

53. Papafitsoros, K.; Adam, L.; Cermák, V.; Picek, L. SeaTurtleID. 2022. Available online: ˇ https://www.kaggle.com/datasets/ wildlifedatasets/seaturtleidheads (accessed on 3 November 2024).

54. Watch, W.T.; Conservation, L.O. Turtle Recall: Conservation Challenge. 2022. Available online: https://zindi.africa/competitions/ turtle-recall-conservation-challenge/data (accessed on 3 November 2024).

55. Trotter, C.; Atkinson, G.; Sharpe, M.; Richardson, K.; McGough, A.S.; Wright, N.; Burville, B.; Berggren, P. The Northumberland Dolphin Dataset 2020; Newcastle University: Newcastle upon Tyne, UK, 2020. Available online: https://doi.org/10.25405/data. ncl.c.4982342 (accessed on 3 November 2024).

56. Humpback Whale Identification, Kaggle. 2018. Available online: https://kaggle.com/competitions/humpback-whaleidentification (accessed on 3 November 2024).

57. Khan, C.B.; Shashank; Kan, W. Right Whale Recognition, Kaggle. 2015. Available online: https://kaggle.com/competitions/ noaa-right-whale-recognition (accessed on 3 November 2024).

58. Holmberg, J.; Norman, B.; Arzoumanian, Z. Whale Shark ID. info@wildme.org. 2020. Available online: https://lila.science/ datasets/whale-shark-id (accessed on 3 November 2024).

59. Gao, J.; Burghardt, T.; Andrew, W.; Dowsey, A.W.; Campbell, N.W. Cows2021. 2021. Available online: https://doi.org/10.5523/ bris.4vnrca7qw1642qlwxjadp87h7 (accessed on 3 November 2024).

60. Andrew, W.; Gao, J.; Mullan, S.; Campbell, N.; Dowsey, A.W.; Burghardt, T. OpenCows2020. 2020. Available online: https: //doi.org/10.5523/bris.10m32xl88x2b61zlkkgz3fml17 (accessed on 3 November 2024).

61. Andrew, W.; Greatwood, C.; Burghardt, T. AerialCattle2017, University of Bristol. 2017. Available online: https://doi.org/10.552 3/bris.3owflku95bxsx24643cybxu3qh (accessed on 3 November 2024).

62. Andrew, W.; Greatwood, C.; Burghardt, T. FriesianCattle2017, University of Bristol. 2017. Available online: https://doi.org/10.5 523/bris.2yizcfbkuv4352pzc32n54371r (accessed on 3 November 2024).

63. Andrew, W.; Hannuna, S.; Campbell, N.; Burghardt, T. FriesianCattle2015, University of Bristol. 2016. Available online: https://doi.org/10.5523/bris.wurzq71kfm561ljahbwjhx9n3 (accessed on 3 November 2024).

64. Kern, D.; Schiele, T.; Klauck, U.; Ingabire, W. Chicks4FreeID. 2024. Available online: https://huggingface.co/datasets/dariakern/ Chicks4FreeID (accessed on 3 November 2024).

65. Lu, W.; Zhao, Y.; Wang, J.; Zheng, Z.; Feng, L.; Tang, J. MammalClub. 2023. Available online: https://github.com/WJ-0425 /MammalClub (accessed on 3 November 2024).

66. He, Z.; Qian, J.; Yan, D.; Wang, C.; Xin, Y. Animal reidentification Algorithm for Posture Diversity. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [CrossRef]

67. He, Z. Multi-Pose Dog Dataset. 2023. Mendeley Data, V1. Available online: https://data.mendeley.com/datasets/v5j6m8dzhv/1 (accessed on 3 November 2024).

68. Dataset for: PolarBearVidID: A Video-Based Re-Identification Benchmark Dataset for Polar Bears. Available online: https://zenodo. org/records/7564529 (accessed on 3 November 2024).

69. Kuncheva, L.I.; Williams, F.; Hennessey, S.L.; Rodríguez, J.J. A Benchmark Database for Animal reidentification and Tracking. In Proceedings of the 2022 IEEE 5th International Conference on Image Processing Applications and Systems (IPAS), Genova, Italy, 5–7 December 2022; Volume 5, pp. 1–6. [CrossRef]

70. Trust, B.P.C. Hyiena ID. Panthera Pardus CSV Custom Export. Retrieved from African Carnivore Wildbook 2022-04-28. 2022. Available online: https://lila.science/datasets/hyena-id-2022/ (accessed on 3 November 2024).

71. Trust, B.P.C. Leopard ID. Panthera Pardus CSV Custom Export. Retrieved from African Carnivore Wildbook 2022-04-28. 2022. Available online: https://lila.science/datasets/leopard-id-2022/ (accessed on 3 November 2024).

72. Nepovinnykh, E.; Eerola, T.; Biard, V.; Mutka, P.; Niemi, M.; Kunnasranta, M.; Kälviäinen, H. SealID: Saimaa Ringed Seal reidentification Dataset. Sensors 2022, 22, 11. [CrossRef]

73. Papafitsoros, K.; Adam, L.; Cermák, V.; Picek, L. SeaTurtleID: A novel long-span dataset highlighting the importance of ˇ timestamps in wildlife re-identification. arXiv 2022, arXiv:2211.10307. [CrossRef]

74. Li, S.; Fu, L.; Sun, Y.; Mu, Y.; Chen, L.; Li, J.; Gong, H. Individual dairy cow identification based on lightweight convolutional neural network. PLoS ONE 2021, 16, 13. [CrossRef]

75. Miele, V.; Dussert, G.; Spataro, B.; Chamaillé-Jammes, S.; Allainé, D.; Bonenfant, C. Revisiting animal photo-identification using deep metric learning and network analysis. Methods Ecol. Evol. 2021, 12, 863–873. [CrossRef]

76. Miele, V.; Dussert, G.; Spataro, B.; Chamaillé-Jammes, S.; Allainé, D.; Bonenfant, C. Giraffe Dataset. 2020. Available online: https://plmlab.math.cnrs.fr/vmiele/animal-reid/ (accessed on 3 November 2024).

77. Wang, L.; Ding, R.; Zhai, Y.; Zhang, Q.; Tang, W.; Zheng, N.; Hua, G. iPanda-50. 2021. Available online: https://github.com/ iPandaDateset/iPanda-50 (accessed on 3 November 2024).

78. Haurum, J.B.; Karpova, A.; Pedersen, M.; Bengtson, S.H.; Moeslund, T.B. AAU Zebrafish Re-Identification Dataset. 2020. Available online: https://www.kaggle.com/datasets/aalborguniversity/aau-zebrafish-reid (accessed on 3 November 2024).

79. Guo, S.; Xu, P.; Miao, Q.; Shao, G.; Chapman, C.A.; Chen, X.; He, G.; Fang, D.; Zhang, H.; Sun, Y.; et al. AFD. Mendeley Data, Version 2. 2020. Available online: https://doi.org/10.17632/z3x59pv4bz.2 (accessed on 3 November 2024).

80. Li, S.; Li, J.; Tang, H.; Qian, R.; Lin, W. ATRW (Amur Tiger Re-identification in the Wild). 2020. Available online: https://lila.science/ datasets/atrw (accessed on 3 November 2024).

81. Dlamini, N.; Zyl, T.L.v. Lion Face Dataset. Mara Masia Project, Kenya. 2020. Available online: https://github.com/tvanzyl/wildlife_ reidentification/ (accessed on 3 November 2024).

82. Trotter, C.; Atkinson, G.; Sharpe, M.; Richardson, K.; McGough, A.S.; Wright, N.; Burville, B.; Berggren, P. NDD20: A large-scale few-shot dolphin dataset for coarse and fine-grained categorisation. arXiv 2020, arXiv:2005.13359.

83. Dlamini, N.; Zyl, T.L.v. Nyala Dataset. South African Nature Reserves. 2020. Available online: https://github.com/tvanzyl/ wildlife_reidentification/ (accessed on 3 November 2024).

84. Ferreira, A.C.; Silva, L.R.; Renna, F.; Brandl, H.B.; Renoult, J.P.; Farine, D.R.; Covas, R.; Doutrelant, C. Deep learning-based methods for individual recognition in small birds. Methods Ecol. Evol. 2020, 11, 1072–1085. [CrossRef]

85. Mougeot, G.; Li, D.; Jia, S. Dog Face Dataset. 2019. Available online: https://github.com/GuillaumeMougeot/DogFaceNet (accessed on 3 November 2024).

86. Lin, T.Y.; Kuo, Y.F. Cat Individual Images. 2018. Available online: https://www.kaggle.com/datasets/timost1234/cat-individuals (accessed on 3 November 2024).

87. Schneider, J.; Murali, N.; Taylor, G.; Levine, J. Can Drosophila melanogaster tell who’s who? PLoS ONE 2018, 13, e0205043. [CrossRef]

88. Dataset for: Can Drosophila Melanogaster Tell Who’s Who? Available online: https://borealisdata.ca/dataset.xhtml? persistentId=doi:10.5683/SP2/JP4WDF (accessed on 3 November 2024).

89. Witham, C.L. MacaqueFaces. 2018. Available online: https://github.com/clwitham/MacaqueFaces (accessed on 3 November 2024).

90. Parham, J.; Crall, J.; Stewart, C.; Berger-Wolf, T.; Rubenstein, D. Great Zebra and Giraffe Count ID. info@wildme.org. 2017. Available online: https://lila.science/datasets/great-zebra-giraffe-id (accessed on 3 November 2024).

91. Freytag, A.; Rodner, E.; Simon, M.; Loos, A.; Kühl, H.S.; Denzler, J. Chimpanzee Faces in the Wild. Acknowledgements: Tobias Deschner, Laura Aporius, Karin Bahrke, Zoo Leipzig. 2016. Available online: https://github.com/cvjena/chimpanzee_faces (accessed on 3 November 2024).

92. Holmberg, J.; Norman, B.; Arzoumanian, Z. Estimating population size, structure, and residency time for whale sharks Rhincodon typus through collaborative photo-identification. Endanger. Species Res. 2009, 7, 39–53. [CrossRef]

93. Labelbox. “Labelbox”. 2024 Available online: https://labelbox.com (accessed on 3 November 2024).

94. Cubuk, E.D.; Zoph, B.; Shlens, J.; Le, Q. RandAugment: Practical Automated Data Augmentation with a Reduced Search Space. In Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 18613–18624.

95. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. 2021. Available online: https://openreview.net/forum?id=YicbFdNTTy (accessed on 8 May 2012).

96. TorchVision: PyTorch’s Computer Vision Library. 2016. Available online: https://github.com/pytorch/vision (accessed on 3 November 2024).

97. Lightly (Software Version 1.5.2). 2020. Available online: https://github.com/lightly-ai/lightly/blob/master/CITATION.cff (accessed on 3 November 2024).

98. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 9992–10002. [CrossRef]

99. Wightman, R. PyTorch Image Models. 2019. Available online: https://github.com/rwightman/pytorch-image-models (accessed on 3 November 2024).

100. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. ArcFace: Additive Angular Margin Loss for Deep Face Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–19 June 2019.

101. Cermák, V.; Picek, L.; Adam, L.; Papafitsoros, K. MegaDescriptor-L-384. 2024. Available online: ˇ https://huggingface.co/BVRA/ MegaDescriptor-L-384 (accessed on 3 November 2024).

102. Oquab, M.; Darcet, T.; Moutakanni, T.; Vo, H.; Szafraniec, M.; Khalidov, V.; Fernandez, P.; Haziza, D.; Massa, F.; El-Nouby, A.; et al. DINOv2: Learning Robust Visual Features without Supervision. arXiv 2024, arXiv:2304.07193.

103. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning Transferable Visual Models From Natural Language Supervision. PMLR 2021, 139, 8748–8763.

104. Wu, Z.; Xiong, Y.; Yu, S.X.; Lin, D. Unsupervised Feature Learning via Non-parametric Instance Discrimination. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3733–3742. [CrossRef]

105. Detlefsen, N.S.; Borovec, J.; Schock, J.; Jha, A.H.; Koker, T.; Liello, L.D.; Stancl, D.; Quan, C.; Grechkin, M.; Falcon, W. TorchMetrics— Measuring Reproducibility in PyTorch. J. Open Source Softw. 2022, 7, 4101. [CrossRef]

106. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the 2nd International Conference on Knowledge Discovery and Data Mining (KDD), Portland, OR, USA, 2–4 August 1996; pp. 226–231.

107. Schubert, E.; Sander, J.; Ester, M.; Kriegel, H.P.; Xu, X. DBSCAN revisited, revisited: Why and how you should (still) use DBSCAN. Acm Trans. Database Syst. (Tods) 2017, 42, 19:1–19:21. [CrossRef]

108. Nielsen, F. Introduction to HPC with MPI for Data Science; Hierarchical Clustering; Springer: Berlin/Heidelberg, Germany, 2016; Chapter 8, pp. 195–211.

Related topics:
Authors:
Daria Kern
Recommend
Comment
Share
Home
Recommend
Comment
Share
Profile picture
Would you like to discuss another topic? Create a new post to engage with experts in the community.
Featured users in Poultry Industry
Lorena Ramos
Lorena Ramos
Cargill
United States
Kendra Waldbusser
Kendra Waldbusser
Pilgrim´s
United States
Carolina Hall
Carolina Hall
Trouw Nutrition
United States
Karen Christensen
Karen Christensen
Tyson
Tyson
PhD, senior director of animal welfare at Tyson Foods
United States
Nelson Ward
Nelson Ward
dsm-Firmenich
Manager, New Business Development at DSM Nutritional Products Inc.
United States