@Override protected void onCreate(Bundle savedInstanceState) { super.onCreate(savedInstanceState); setContentView(R.layout.activity_region_list); Intent i = getIntent(); String state = i.getExtras().getString(Constants.SELECTED_STATE); String openFile = checkState(state); // declare reader class and input stream ReadFile fileReader; InputStream input = null; // try to read categories file fileReader = new ReadFile(); cities = new ArrayList<String>(); searchCities = new ArrayList<String>(); try { input = getAssets().open(openFile + ".txt"); cities = fileReader.OpenFile(input); input = getAssets().open(openFile + "links.txt"); searchCities = fileReader.OpenFile(input); } catch (IOException e) { // TODO Auto-generated catch block e.printStackTrace(); } ArrayAdapter<String> adapter = new ArrayAdapter<String>(this, android.R.layout.simple_list_item_1, cities); setListAdapter(adapter); }
@Override public void readFile() { // TODO Auto-generated method stub ReadFile temp = new ReadFile(mfileName); mworkBook = temp.readFile(mworkBook); msheetCount = mworkBook.getNumberOfSheets(); getDataFromSheet(); }
private static String readCode(final String filename) { final ReadFile reader = new ReadFile(filename); try { return reader.read(); } catch (Exception ex) { ex.printStackTrace(System.out); throw new RuntimeException(ex); } }
/* -- This is a helper methid to run the morph files */ private static void runMorphDataSet() throws Exception { String morph_directory = "../../thesis-datasets/morph/"; // directory where all the morph code is stored File d = new File(morph_directory); // get all the files from a directory File[] fList = d.listFiles(); List<String> dir_list = new ArrayList<String>(); for (File file : fList) { if (file.isDirectory()) { dir_list.add(file.getName()); } } for (String dir : dir_list) { directory = morph_directory + dir + "/"; System.out.println("Running TDDD " + directory); ReadFile.readFile(directory, fileList); // read the two files System.out.println(fileList.get(0) + " " + fileList.get(1)); preliminaryStep(directory); startCDC(); fileList.clear(); fileArray.clear(); hashed_File_List.clear(); } }
/** * Contains the 'main' method. It issues calls to various methods to perform the following * functions: - Read data from input file - Create a visualisation of list of states - Create a * tree-map of the data - Create a mashup for the data */ public static void main(String[] args) { ReadFile.run(); StateMenu.run(); TreeMap.run(); MakeSummaryTable.make(); Mashup.display(MakeSummaryTable.summary); }
/* -- This is a helper method run datasets such as emacs, gcc etc */ private static void runOtherDataSets() throws Exception { System.out.println("Running tddd " + directory); ReadFile.readFile(directory, fileList); // read the two files System.out.println(fileList.get(0) + " " + fileList.get(1)); preliminaryStep(directory); startCDC(); }
/** * This method stores the first column of the file(the data type) on one array and the second * column (the data value) on another array * * @param contents the contents of the text file */ public void process_Qfile(String contents) { String delimiter = "-------"; // each attribute separated by -- // Get all the rows (string array of them) String[] contents_split; contents_split = reader.split(contents, "\n"); String[] line_split; data_type_list = new String[contents_split.length]; // create data type array data_value_list = new String[contents_split.length]; // create data value array for (int i = 0; i < contents_split.length; i++) { line_split = reader.split(contents_split[i], delimiter); // splits the contents of the file in columns data_type_list[i] = line_split[0]; // store the first column containing the data type data_value_list[i] = line_split[1]; // store the second column containing the data value } }
// this method basically will chop up the blocks and get their frequencies private static void getBlockFrequency() throws Exception { directory = "../../thesis-datasets/morph_file_100MB/"; ReadFile.readFile(directory, fileList); // read the two files HashMap<Integer, Integer> blockFreq = new HashMap<Integer, Integer>(); // this stores the block in the map along there frequencies int start = 0; // start of the sliding window int end = start + window - 1; // ending boundary preliminaryStep(directory); // System.out.println("Choping the document TDDD " + fileList.get(0)); long[] divisorArray = { 1000 }; // run the frequency code for these divisor values (AKA expected block Size) for (long i : divisorArray) { long divisor1 = i; long divisor2 = i / 2; long divisor3 = i / 4; long remainder = 7; long minBoundary = min_multiplier * i; long maxBoundary = max_multiplier * i; // System.out.println("Running Likelihood for " + i + " " + divisor2 + " " + divisor3); int totalBlocks = chopDocument( fileArray.get(0), hashed_File_List.get(0), divisor1, divisor2, divisor3, remainder, minBoundary, maxBoundary, blockFreq); // now output the block sizes, along with there frequencies and probilities for (Map.Entry<Integer, Integer> tuple : blockFreq.entrySet()) { // output the block freq double prob = (double) tuple.getValue() / (double) totalBlocks; System.out.println(tuple.getKey() + " " + tuple.getValue() + " " + prob); } blockFreq.clear(); } }
/** * @param rankedList: a ranked list of turkers ID, the length should be 51 * @return: the min number of turkders that guarantees all sentences are covered */ public static int pruneRankedTurkersList(ArrayList<String> rankedList) { FormatTurkerFileInput tmp = new FormatTurkerFileInput(); // read the input from file ArrayList<Integer> tmpSentsList = new ArrayList<Integer>(); init(); // set up the initial table for sentences // loop from the last element for (int turkerRankIdx = totalTurkNum - 1; turkerRankIdx >= 0; --turkerRankIdx) { // get all the sentences' ID that a single person translated tmp = ReadFile.turkerFileInput(rankedList.get(turkerRankIdx)); tmpSentsList = tmp.getTranslatedSentNum(); // for each sentence ID, modify the corresponding table value Iterator<Integer> iter = tmpSentsList.iterator(); while (iter.hasNext()) { int tmpSent = iter.next(); sentenceTurkNum.put(tmpSent, sentenceTurkNum.get(tmpSent) - 1); if (sentenceTurkNum.get(tmpSent) == 0) { // when any value reaches zero, return the number of left turkers return (turkerRankIdx + 1); } } } return 0; }
/* - This method is used has a helper method to run the algo for the archive dataset - Note the archive set has multiple directories ( one for each url ) - So Read all of the directories in first and for each directory run the code */ private static void runArchiveSet() throws Exception { System.out.println("Running TDDD archive"); directory = "../../thesis-datasets/datasets2/"; File file = new File(directory); String[] directory_list = file.list( new FilenameFilter() { @Override public boolean accept(File current, String name) { return new File(current, name).isDirectory(); // make sure its a directory } }); int totalRuns = 0; // used to avg the runs in the end int total_iter_count = 0; // this is used check how many times we will iterate through the data so we can make an // array of that size for (int i = startBoundary; i <= endBoundary; i += increment) total_iter_count++; // System.out.println(Arrays.toString(directory_list)); int sets = 0; // make the arrays to hold the respecitve info for the different verions\ // run it simulateounsly to speed the from the program! double[] block_size_list_last_year = new double[total_iter_count]; double[] ratio_size_list_last_year = new double[total_iter_count]; double[] block_size_list_six_month = new double[total_iter_count]; double[] ratio_size_list__six_month = new double[total_iter_count]; double[] block_size_list_two_year = new double[total_iter_count]; double[] ratio_size_list_two_year = new double[total_iter_count]; int current = 0; int six_month = 2; int last_year = 1; int two_year = 3; // loop through and run the cdc for each directory for (String dir : directory_list) { ReadFile.readFile(directory + dir, fileList); // read all the files in this directory preliminaryStep(directory + dir + "/"); // call the preliminaryStep on all the files totalRuns++; totalSize = fileArray.get(current) .length; // get the length of the file we will be running it against! // run it against six month startCDC( block_size_list_six_month, ratio_size_list__six_month, fileArray.get(current), fileArray.get(six_month), hashed_File_List.get(current), hashed_File_List.get(six_month)); // run it against last year startCDC( block_size_list_last_year, ratio_size_list_last_year, fileArray.get(current), fileArray.get(last_year), hashed_File_List.get(current), hashed_File_List.get(last_year)); // run it against 2 startCDC( block_size_list_two_year, ratio_size_list_two_year, fileArray.get(current), fileArray.get(two_year), hashed_File_List.get(current), hashed_File_List.get(two_year)); // // clear the fileList and hashed_file_list array fileArray.clear(); hashed_File_List.clear(); fileList.clear(); // if (Double.isNaN(ratio_size_list[0])){ // System.out.println(sets+" "+Arrays.toString(ratio_size_list)); // test = true; // break; // } if (sets % 200 == 0) System.out.println(sets); ++sets; } // end of directory list for loop // now output the avged value for all the runs // System.out.println(Arrays.toString(ratio_size_list)); System.out.println("Printing six_month"); int index = 0; for (int i = startBoundary; i <= endBoundary; i += increment) { // avg out the outputs double blockSize = block_size_list_six_month[index] / (double) totalRuns; double ratio = ratio_size_list__six_month[index] / (double) totalRuns; System.out.println(i + " " + i / 2 + 1 + " " + i / 4 + 1 + " " + blockSize + " " + ratio); index++; } System.out.println("Printing last year"); index = 0; for (int i = startBoundary; i <= endBoundary; i += increment) { double blockSize = block_size_list_last_year[index] / (double) totalRuns; double ratio = ratio_size_list_last_year[index] / (double) totalRuns; System.out.println(i + " " + blockSize + " " + ratio); index++; } System.out.println("Printing two year"); index = 0; for (int i = startBoundary; i <= endBoundary; i += increment) { double blockSize = block_size_list_two_year[index] / (double) totalRuns; double ratio = ratio_size_list_two_year[index] / (double) totalRuns; System.out.println(i + " " + blockSize + " " + ratio); index++; } }
@Before public void setUp() throws Exception { ReadFile readObj = new ReadFile("csvFiles/sample.csv"); data = readObj.readContent(); }
public static ArrayList<Transaction> readDB(String un) { ArrayList<String[]> splitFile = ReadFile.readInUser(un + "/Trans.csv"); return readTransactionImports(splitFile); }
public static void main(String[] args) throws IOException { String file_name = "Integers.java"; ReadFile file = new ReadFile(file_name); file.OpenFile(); }
public boolean getByteInCallback(final ByteInCallback _callback, final int _length) { return ReadFile.getRaw(getByteInStream(), _callback, _length); }
public boolean getStringInCallback(final StringInCallback _callback, final int _length) { return ReadFile.getString(getStringInStream(), _callback, _length); }
/* -- This is a helper method to run the periodic dataset basically */ private static void runPeriodic() throws Exception { System.out.println("Running TDDD Periodic"); // this is alll the directories we will be running int arr[] = {10, 15, 20, 25, 30}; // this is the input number we will be running on // this is the base of the two files // these two are directories, we will concanate with the numbers to get the full dir name String base_old_file = "../../thesis-datasets/input_"; String base_new_file = "../../thesis-datasets/periodic_"; int total_iter_count = 0; // this is used check how many times we will iterate through the data so we can make an // array of that size for (int i = startBoundary; i <= endBoundary; i += increment) total_iter_count++; for (int dir_num : arr) { // set up our directories String old_file_dir = base_old_file + dir_num + "/"; String new_file_dir = base_new_file + dir_num + "/"; System.out.println(old_file_dir); // read all the files in these two directories in sorted order ArrayList<String> old_file_list = new ArrayList<String>(); ArrayList<String> new_file_list = new ArrayList<String>(); ReadFile.readFile(old_file_dir, old_file_list); ReadFile.readFile(new_file_dir, new_file_list); // used to store all the runnings for the periodic data double[] block_size_list = new double[total_iter_count]; double[] ratio_size_list = new double[total_iter_count]; int totalRuns = 0; for (int i = 0; i < old_file_list.size(); ++i) { // System.out.println(old_file_list.get(i) + " " + new_file_list.get(i)); String[] s1 = old_file_list.get(i).split("_"); String[] s2 = new_file_list.get(i).split("_"); // input file should corrospond to the output file if (!s1[1].equals(s2[1]) || !s1[2].equals(s2[2])) System.out.println("We got a huge problem"); // basically same code as in the prelinaryStep method, but we need to modify it for // perdiodic files int start = 0; // start of the sliding window int end = start + window - 1; // ending boundary // we cant call preliminary function, so hash the two files individually // System.out.println("preliminaryStep " + fileList.get(i)); Path p = Paths.get( old_file_dir + old_file_list.get( i)); // read the old file ( the one which we will be using as the base // comparason) Path p2 = Paths.get( new_file_dir + new_file_list.get( i)); // read the old file ( the one which we will be using as the base // comparason) byte[] old_file = Files.readAllBytes(p); // read the file in bytes byte[] new_file = Files.readAllBytes(p2); // System.out.println(array.length); ArrayList<Long> old_file_hashes = new ArrayList<Long>(); // make a new arrayList for this document ArrayList<Long> new_file_hashes = new ArrayList<Long>(); // make a new arrayList for this document HashDocument.hashDocument( old_file, old_file_hashes, start, end); // this hashes the entire document using the window and stores itto md5hashes // array HashDocument.hashDocument( new_file, new_file_hashes, start, end); // this hashes the entire document using the window and stores itto md5hashes // array // now call the startCdc method totalSize = new_file.length; // this is the length of the file startCDC( block_size_list, ratio_size_list, new_file, old_file, new_file_hashes, old_file_hashes); if (totalRuns % 10 == 0) System.out.println(totalRuns); totalRuns++; } // now output the results System.out.println("File dir = " + dir_num + " totalRuns = " + totalRuns); int index = 0; for (int i = startBoundary; i <= endBoundary; i += increment) { // avg out the outputs double blockSize = block_size_list[index] / (double) totalRuns; double ratio = ratio_size_list[index] / (double) totalRuns; System.out.println(i + " " + blockSize + " " + ratio); index++; } // now each index matches the corrosponding file } } // end of methid
public Questions(String currentUser, int week) { // current user this.currentUser = currentUser; // set week this.week = week; // set title setTitle("Questions"); // background defaultBg = BackgroundFactory.createLinearGradientBackground( Color.WHITE, Color.LIGHTSKYBLUE, Color.WHITE, Color.GRAY); // main manager mainManager = (VerticalFieldManager) getMainManager(); mainManager.setBackground(defaultBg); // create buttons goBack = new ButtonField("Go back", ButtonField.FIELD_HCENTER); // create reader object reader = new ReadFile(); // read the question file and return its contents test = reader.readTextFile("Questions2.txt"); // process the information process_Qfile(test); // create and set the question objects CreateAndSetQuestionObj(); // create Question list questions_list = CreateQList(); // drop-down list questionsList = new ObjectChoiceField("This week's Questions: ", questions_list); // set margins questionsList.setMargin(20, 0, 0, 0); goBack.setMargin(30, 0, 0, 0); // add buttons to the listener questionsList.setChangeListener(this); goBack.setChangeListener(this); // create image and overwrite the paint method of verticalFieldManager // backgroundImage = Bitmap.getBitmapResource("beach4 copy1.jpg"); backgroundImage = Bitmap.getBitmapResource("beach5 copy copy.jpg"); mainManager = new VerticalFieldManager( VerticalFieldManager.USE_ALL_WIDTH | VerticalFieldManager.USE_ALL_HEIGHT) { public void paint(Graphics graphics) { graphics.drawBitmap(0, 0, 480, 510, backgroundImage, 0, 0); super.paint(graphics); } }; this.add(mainManager); // add components mainManager.add(questionsList); mainManager.add(goBack); // decide which panel should display displayAnswer(); }
/** @return */ public static double calcScoreForMinTurks( ArrayList<String> rankedList, boolean doPrint, int mode) { // mode 1 for not prune, mode 2 for prune int listLen = 0; ArrayList<String> prunedList = new ArrayList<String>(); if (mode == 1) { prunedList = rankedList; listLen = prunedList.size(); } if (mode == 2) { listLen = pruneRankedTurkersList(rankedList); prunedList = pruneList(listLen, rankedList); } if (mode != 1 && mode != 2) { System.out.println("No such mode for calcScoreForMinTurks."); return -1; } final int sntnsNum = 1683; // 1683 files in total int sentsLeft = sntnsNum; int sntnsIdx = 1; double scoreSum = 0; double scoreThis, tmp; double[] tmp1 = new double[4]; while (sntnsIdx <= sntnsNum) { // gather data FormatSentFileInput oneSentData = ReadFile.sentFileInput(sntnsIdx); String[] lst = oneSentData.getRefList(); String[] lst1 = new String[3]; HashMap<String, String> dct = oneSentData.getTransDict(); // initialize scoreThis = 0; // traverse all the translated sentence for one Urdu sentence // and find the highest score Iterator<Entry<String, String>> iter = dct.entrySet().iterator(); while (iter.hasNext()) { Map.Entry<String, String> entry = (Map.Entry<String, String>) iter.next(); String key = entry.getKey(); String val = entry.getValue(); // the bleu score will be calculated and added only when the turker is in the list if (prunedList.contains(key)) { lst1[0] = lst[0]; lst1[1] = lst[1]; lst1[2] = lst[2]; tmp1[3] = BLEU.computeSentenceBleu(lst1, val); // compute bleu score lst1[2] = lst[3]; tmp1[2] = BLEU.computeSentenceBleu(lst1, val); // compute bleu score lst1[1] = lst[2]; tmp1[1] = BLEU.computeSentenceBleu(lst1, val); // compute bleu score lst1[0] = lst[1]; tmp1[0] = BLEU.computeSentenceBleu(lst1, val); // compute bleu score tmp = (tmp1[0] + tmp1[1] + tmp1[2] + tmp1[3]) / 4; scoreThis = (tmp > scoreThis) ? tmp : scoreThis; // update scoreThis to the highest score } } if (scoreThis == 0) { --sentsLeft; } scoreSum += scoreThis; sntnsIdx++; } if (doPrint) { System.out.println("Sentences left: " + sentsLeft); System.out.println("Turker left: " + listLen); System.out.println("Best average BLEU score: " + scoreSum / sentsLeft); } return scoreSum / sentsLeft; }