Great Deal! Get Instant $10 FREE in Account on First Order + 10% Cashback on Every Order Order Now

Homework 5 (20 points ) For this homework, you can work in a team of 2 members maximum . Each team should turn in one program. Make sure that all of the team members' names are listed at the top of...

1 answer below »
Homework 5

(20 points )

For this homework, you can work in ateam of 2 members maximum. Each team should turn in one program. Make sure that all of the team members' names are listed at the top of the program. If you are the one doing the turn in for the team, make sure that you know the access code of your partner.

You may also work alone if you wish.

In this assignment, you will practice doing the following:

  • designing classes (using an interface),
  • using 2D arrays

Digital image processing has completely revolutionized the way images are created and used in news photography, publishing, commercial art, marketing, and even in some of the fine arts. Adobe’s Photoshop program has become so ubiquitous that it has even become a verb - “this picture is a mess, I need to Photoshop it”. In this assignment, you’ll implement some of the core image transformation algorithms used by image processing programs like Photoshop.

Image Representation

A digital image is a rectangular array of pixel objects. Each pixel contains three integer values that range from 0 to 255, one integer each for the red, green, and blue components of the pixel, in that order. The larger a number, the brighter that color appears in the pixel. So a pixel with values (0,0,0) is black, (255,255,255) is white, (255,0,0) is pure red, (0,0,255) is pure blue, and so forth.

We can represent a Pixel as a simple Java object containing three int instance variables and, for convenience, a constructor to create a new pixel given its rgb component values. As we did with the rainfall example in lecture, we'll treat these as simple data objects and allow direct references to their fields.

/** Representation of one pixel */
public class Pixel {
public int red; // rgb values in the range 0 to 255
public int green;
public int blue;
/** Construct a new pixel with specified rgb values */
public Pixel(int r, int g, int b) {
this.red = r;
this.green = g;
this.blue = b;
}
}
An image is represented by an instance of class PixelImage. This class contains the methodsgetData()andsetData()to retrieve 2-dimensional arrays of Pixel objects representing the pixels of the image. You can also use the methodsgetHeight()andgetWidth()to get the height and width of the image.

The Application

We have implemented a PhotoShop-like Java application, called SnapShop. The application knows how to load image files, and provides all the user interface objects you'll need to apply your filters to the image. You will need not implement anything in the SnapShop class, but create a few new classes. The file loader of the SnapShop class expects filenames to be fully specified, that is, you must say something likec:\directory\image.jpg. Normal (forward) slashes also work:c:/directory/image.jpg. To save you some work, we've provided a shortcut so you don't need to always retype the file name; details below. To run the application, select the class SnapShop within eclipse and click the Run button.

We have also provided an interface class Filter. Remember that an interface simply specifies the methods another class must implement, and cannot be used to make objects itself. You will be writing classes that implement this Filter interface, one for each transformation you write. Each class implementing the Filter interface must have a method calledfilter(), which takes a PixelImage as an argument. The method then applies a transformation to the data in the image. As an example, we have included the FlipHorizontalFilter class, which flips an image horizontally.

You will need some way to tell our SnapShop class which filters you have implemented. So we've provided a class called SnapShopConfiguration, with a single method,configure(). In this method, you can call methods for the SnapShop object. The two methods you'll be interested in areaddFilter(), which creates a button in the application to apply your filter, andsetDefaultFilename(), which lets you specify a default path or filename for the file loader, to aid you in testing. For each filter that you create, there should be a call toaddFilter()in theconfigure()method.

(A note onsetDefaultFilename(). Windows path names have backslashes (\) in them. To specify this in a string in a Java source program, you need to put two backslashes. For example, the Java string forc:\directory\image.jpgwould be"c:\\directory\\image.jpg".)
Here are the files you need to download to get started:SnapShop.java,SnapShopConfiguration.java,Pixel.java,PixelImage.java,Filter.java,FlipHorizontalFilter.java. Here are a couple of images you can use to test your code:billg.jpg,seattle.jpg,teddy_bear.jpg. If you use other images as test images, be sure that they aren't much bigger than seattle.jpg. Your program will take a long time to process a large image.

Simple Transformations

There are two kinds of transformations that you are required to implement. The simple transformations can be implemented by replacing each Pixel in the existing image with the updated one. The more complex 3x3 transformations require creating a new array of Pixels with the transformed image, then updating the image instance variable to refer to the new array once it is completely initialized.

The first three transformations you should implement flip the image horizontally and vertically, and transform the image into a photographic negative of itself (that is, you should create a flipHorizontalFilter, flipVerticalFilter, and NegativeFilter class). We have implemented flipHorizontalFilter for you.

The first two require a simple rearrangement of the pixels that reverses the order of rows or columns in the image. The negate transformation is done by replacing each Pixel in the image with a new Pixel whose rgb values are calculated by subtracting the original rgb values from 255. These subtractions are done individually for each of the red, green, and blue colors.

These transformations can be performed by modifying the image array of Pixels directly. You should do these first to get a better idea of how the image is represented and what happens when you modify the Pixels. You should make every effort to get this far before the end of week 8 (2/28). That will ensure that you’ve made good progress on this assignment, or at least know what you need to clear up in discussions during lecture.

Notes:

  • You can assume that the image array is rectangular, i.e., all rows have the same length.
  • All of these simple transformations are their own inverse. If you repeat any of these transformations twice in a row, you should get the original image back.
  • You should use relatively small image files for testing. The program will work fine with large images, but there can be a significant delay while the transformed image is created if the image is large.

3x3 Transformations

Once you’ve got the simple transformations working, you should implement this next set, which includes Gaussian blur, Laplacian, Unsharp Masking, and Edgy. All of these transformations are based on the following idea: each pixel in the transformed image is calculated from the values of the original pixel and its immediate neighbors, i.e., the 3x3 array of pixels centered on the old pixel whose new value we are trying to calculate. The new rgb values can be obtained by calculating a weighted average; the median, minimum, or maximum; or something else. As with the negate transformation, the calculations are carried out independently for each color, i.e., the new red value for a pixel is obtained from the old red values, and similarly for red and blue.

The four transformations you should implement all compute the new pixel values as a weighted average of the old ones. The only difference between them is the actual weights that are used. You should be able to add a single method inside class PixelImage to compute a new image using weighted averages, and call it from the methods for the specific transformations with appropriate weights as parameters.You shouldnotneed to repeat the code for calculating weighted averages four times, once in each transformation. The method you add to PixelImage to do the actual calculations can, of course, call additional new methods if it makes sense to break the calculation into smaller pieces.

Here are the weights for the 3x3 transformations you should implement.

Gaussian

1 2 1
2 4 2
1 2 1

After computing the weighted sum, the result must be divided by 16 to scale the numbers back down to the range 0 to 255. The effect is to blur the image.

Laplacian

XXXXXXXXXX
-1 8 -1
XXXXXXXXXX

The neighboring pixel values are subtracted from 8 times the center one, so no scaling is needed. However, you do need to check that the weighted average is between 0 and 255. If it is less than 0, replace the calculated value with 0 (i.e., the new value is the maximum of 0 and the calculated value). If it is greater than 255, then replace the calculated value with 255. This transformation detects and highlights edges.

Unsharp masking

XXXXXXXXXX
XXXXXXXXXX
XXXXXXXXXX

This transformation is created by multiplying the center pixel and subtracting the Gaussian weighted average. The result must be divided by 16 to scale it back down to the range 0 to 255. As with the Laplacian transformation, check for negative weighted averages or weighted averages greater than 255 (and do the same thing as in the Laplacian case to fix it).

Edgy

XXXXXXXXXX
-1 9 -1
XXXXXXXXXX

This adds the Laplacian weighted average to the original pixel, which sharpens the edges in the image. It does not need scaling, but you need to watch for weighted averages less than 0 or greater than 255.

Notes:
  • The complication with these transformations is that the new value of each pixel depends on the neighboring ones, as well as itself. That means we cannot replace the original pixels with new values before the old values have been used to compute the new values of their neighbors. The simplest way to handle this is to create a new 2D Pixel array the same size as the old image, compute Pixels for the new image and store them in the new array, then change the image instance variable to refer to the new array once it is completed.
  • You should assume the image has at least three rows and columns and you do not need to worry about updating the first and last rows and columns. In other words,only update the interior pixelsthat have neighbors on all four sides. However, every position in the array of Pixels must have refer to a Pixel object; you can't just leave a position in the array uninitialized.
  • Debugging hint: From past experience, we’ve noticed that bugs in the implementation of these transformations tend to produce more spectacular visible effects with the Laplacian weights. You might start with this set of weights when testing your code for the 3x3 transformations.
  • Be sure that your monitor is set to “thousands” or “millions” of colors, which is normally the case on most modern PCs. If you set the monitor to such a high resolution that the color display is set to 256, the colors will be rendered only approximately and it will be hard to see the effects of most of these transformations.

Further explorations (up to 5 points of extra-credit)

Once you have the basic assignment working, turn it in so you’ve got something submitted. Then feel free to experiment with additional transformations. Besides varying the weights, you can try replacing each pixel by the minimum or maximum value in the neighborhood, or the median. Try weights that are not symmetric. Try using a larger neighborhood like 5x5. Or look on the web or in the library (an actual library not the Java library!) for additional things you can do with the images. Some amount of extra credit (up to 5 points) will be awarded for particularly imaginative work.

Here are sometransformation examplesdone with billg.jpg

Report

Write a short report (as a comment) at the beginning of the SnapShopConfiguration file. Describe:

  • what additional filters (besides the three simple transformations and the four 3x3 transformations) you wrote and submitted, if any
  • what works and what doesn’t
  • the surprises or problems you encountered while implementing the transformations.

(-2 points if you don't provide a report)

Your program has to be your own.
Answered 5 days After Feb 28, 2022

Solution

Neetha answered on Mar 06 2022
97 Votes
SOLUTION.PDF

Answer To This Question Is Available To Download

Related Questions & Answers

More Questions »

Submit New Assignment

Copy and Paste Your Assignment Here