Professional Documents
Culture Documents
CS AIML 3A IceSight
CS AIML 3A IceSight
CS AIML 3A IceSight
ON
“IceSight”
Submitted in
Partial Fulfilment of requirements for the Award of Degree
of
Bachelor of Technology
In
Computer Science and Engineering
(Artificial Intelligence & Machine Learning)
By
2. Project Objective
Our project will provide the mapping of glacial areas through satellite images. With this,
change in glacial regions can be observed and will be helpful predicting disasters beforehand
and relative protective measures and rescue operations can be planned.
3. Feasibility Study:
Technical Feasibility:
The project's technical feasibility is high. Cutting-edge technologies like Python, OpenCV,
and TensorFlow, along with the established U-Net architecture, provide robust foundations
for image processing and machine learning tasks. GeoTIFF slicing and numpy file
conversion methods have been successfully implemented, demonstrating the technical
viability of data preprocessing. Additionally, the development environment and tools
required are readily available and accessible.
Operational Feasibility:
The operational feasibility is promising. The project workflow, involving data preprocessing,
model training, and user interaction, is structured to be user-friendly. The slicing and
transformation processes have been optimized for efficiency, allowing for seamless handling
of large datasets. The human-in-the-loop correction system enables easy user interaction,
ensuring accurate results. A user-friendly web interface will facilitate intuitive interactions,
making the system accessible to a wide audience.
Economic Feasibility:
The economic feasibility is favorable. The project predominantly utilizes open-source
software and freely available datasets, reducing upfront costs. While hardware resources for
model training may require initial investment, cloud-based solutions offer cost-effective
alternatives. Furthermore, the potential societal and environmental benefits of improved
glacial monitoring justify any moderate financial investment required.
Schedule Feasibility:
The project timeline is realistic. Based on the outlined methodology, the project is estimated
to be completed within a 6-month timeframe. This includes data collection, preprocessing,
model development, and deployment. The timeline allows for ample iterations and testing
phases to ensure the system's accuracy and reliability. Regular progress assessments and
milestone tracking will be implemented to ensure adherence to the schedule.
Legal Feasibility:
The project aligns with legal requirements and standards. Data collection and usage adhere
to ethical guidelines and respect privacy regulations. All data sources and licenses are
properly documented and attributed. Intellectual property rights and licensing considerations
have been reviewed, ensuring compliance with relevant legal frameworks. The system's
deployment and use will be subject to appropriate terms of service and privacy policies.
Gantt Chart:
4. Methodology/ Planning of work
1.Data Preprocessing:
1.1 Slicing:
In this step, we slice the input GeoTIFF images into smaller 512x512 tiles. Slicing helps
manage large images, and these smaller tiles will be used as input for your deep learning
model.
Alongside these tiles, we also store the corresponding shapefile labels. These labels outline
glacier regions and serve as ground truth for your model.
1.2 Transformation:
To facilitate efficient processing, we convert both the sliced input GeoTIFF images and
shapefile labels into multi-dimensional NumPy .npy files. NumPy files are a suitable format
for handling large datasets and are easily readable by many deep learning frameworks.
1.3. Masking:
The input shapefiles, which represent glacier regions, are transformed into masks. Masks
are essential for use as labels during training.
The transformation involves converting the label data into multi-channel images, where each
channel represents a label class. For example, channel 0 might represent glaciers, channel 1
might represent debris, and so on.
2.Data Postprocessing:
2.1. Filtering:
This step returns the paths for pairs of data (images and labels) that pass specific filter criteria
for a particular channel. This filtering helps select data that is relevant to our training
objectives.
After filtering, the final dataset is organized and saved in three folders: train, test, and
validation. This splitting allows us to create separate datasets for training, testing, and
validation, which are crucial for model development and evaluation.
2.3. Reshuffle:
Shuffling the images and masks in the output directory is a common practice to introduce
randomness and avoid any potential biases in the data order. Randomly shuffling the data helps
ensure that the model does not learn patterns related to the order in which the data was
presented.
2.4. Imputation:
This step involves checking for missing values (NaNs) in the data and replacing them
with zeros. Ensuring there are no missing values in the dataset is important for model stability
and accuracy.
2.5. Normalization:
Finally, we normalize the final dataset based on the calculated means and standard
deviations. Normalization helps ensure that the data has a consistent scale and distribution,
which is important for training deep learning models.
Implement the U-Net architecture using TensorFlow. The U-Net consists of an encoder
network and a decoder network. The encoder down samples the input image to capture high-level
features, while the decoder up samples to produce a segmentation mask. Use appropriate
activation functions (e.g., ReLU) and batch normalization to enhance training stability.
4. Model Training:
Initialize the U-Net model with random weights. Use the Dice loss function as the
optimization criterion. The Dice loss measures the similarity between predicted and ground
truth masks. Train the model using gradient descent or an optimizer like Adam. Monitor
training progress with metrics such as accuracy, loss, and Dice coefficient. Implement early
stopping to prevent overfitting.
5. Human-in-the-Loop Correction:
After initial model training, deploy the model as an interactive web tool. Users can
visualize and interact with the model's segmentation predictions. Allow users to correct any
segmentation errors made by the model by drawing or editing the masks.
5. Tools/Technology Used:
5.1 Minimum Hardware Requirements