|
#!/usr/bin/env ruby |
|
require_relative 'config' |
|
require 'open3' |
|
require 'json' |
|
|
|
|
|
|
|
|
|
|
|
|
|
def copy_task_without_test_output(source_task_json_path, destination_task_json_path) |
|
if source_task_json_path == destination_task_json_path |
|
raise "the paths are supposed to be different. #{source_task_json_path}" |
|
end |
|
json_string = IO.read(source_task_json_path) |
|
json = JSON.parse(json_string) |
|
test_pairs = json['test'] |
|
test_pairs.each do |pair| |
|
pair['output'] = [] |
|
end |
|
File.write(destination_task_json_path, JSON.dump(json)) |
|
end |
|
|
|
|
|
|
|
|
|
def size_from_json_image(rows) |
|
columns_min = 255 |
|
columns_max = 0 |
|
rows.each do |row| |
|
columns_max = [columns_max, row.count].max |
|
columns_min = [columns_min, row.count].min |
|
end |
|
if columns_min != columns_max |
|
raise "the columns are supposed to have the same length. #{task_json_path}" |
|
end |
|
width = columns_min |
|
height = rows.count |
|
"#{width}x#{height}" |
|
end |
|
|
|
def same_size_for_input_and_output_in_task(task_json_path) |
|
json_string = IO.read(task_json_path) |
|
json = JSON.parse(json_string) |
|
sizes_input = [] |
|
sizes_output = [] |
|
json['train'].each do |pair| |
|
sizes_input << size_from_json_image(pair['input']) |
|
sizes_output << size_from_json_image(pair['output']) |
|
end |
|
json['test'].each do |pair| |
|
sizes_input << size_from_json_image(pair['input']) |
|
sizes_output << size_from_json_image(pair['output']) |
|
end |
|
sizes_input == sizes_output |
|
end |
|
|
|
|
|
|
|
|
|
def sizes_from_task(task_json_path) |
|
json_string = IO.read(task_json_path) |
|
json = JSON.parse(json_string) |
|
test_pairs = json['test'] |
|
sizes = [] |
|
test_pairs.each do |pair| |
|
rows = pair['output'] |
|
sizes << size_from_json_image(rows) |
|
end |
|
sizes |
|
end |
|
|
|
|
|
|
|
|
|
def predicted_sizes(json_string) |
|
json = JSON.parse(json_string) |
|
test_pairs = json['test'] |
|
sizes = [] |
|
test_pairs.each do |pair| |
|
dict = pair['output_size'] |
|
width = dict['width'].to_i |
|
height = dict['height'].to_i |
|
sizes << "#{width}x#{height}" |
|
end |
|
sizes |
|
end |
|
|
|
OUTPUT_DIR = File.expand_path("data/arc_size") |
|
TEMP_PATH = File.join(OUTPUT_DIR, 'temp.json') |
|
|
|
LODA_RUST_EXECUTABLE = Config.instance.loda_rust_executable |
|
unless File.executable?(LODA_RUST_EXECUTABLE) |
|
raise "No such file #{LODA_RUST_EXECUTABLE}, cannot run script" |
|
end |
|
|
|
ARC_REPOSITORY_DATA = Config.instance.arc_repository_data |
|
unless File.directory?(ARC_REPOSITORY_DATA) |
|
raise "No such dir #{ARC_REPOSITORY_DATA}, cannot run script" |
|
end |
|
|
|
if File.directory?(OUTPUT_DIR) |
|
raise "The OUTPUT_DIR #{OUTPUT_DIR} already exist. Please delete it manually, and try again." |
|
end |
|
|
|
FileUtils.mkdir_p(OUTPUT_DIR) |
|
unless File.directory?(OUTPUT_DIR) |
|
raise "unable to create dir: #{OUTPUT_DIR}" |
|
end |
|
|
|
count_tasks = 0 |
|
count_ok_predictions = 0 |
|
count_bad_predictions = 0 |
|
count_cannot_predict = 0 |
|
count_other_errors = 0 |
|
count_same_size = 0 |
|
Dir.chdir(ARC_REPOSITORY_DATA) do |
|
paths = Dir.glob("**/*.json") |
|
|
|
|
|
paths = paths.reject { |path| File.basename(path) == 'solution_notXORdinary.json' } |
|
|
|
paths.each_with_index do |path, index| |
|
if index % 100 == 0 |
|
puts "Progress: #{index} of #{paths.count}" |
|
end |
|
|
|
if same_size_for_input_and_output_in_task(path) |
|
count_same_size += 1 |
|
end |
|
|
|
|
|
expected_sizes = sizes_from_task(path) |
|
|
|
|
|
copy_task_without_test_output(path, TEMP_PATH) |
|
|
|
|
|
output_path = File.join(OUTPUT_DIR, path) |
|
output_dirname = File.dirname(output_path) |
|
FileUtils.mkdir_p(output_dirname) |
|
unless File.directory?(output_dirname) |
|
raise "unable to create dir: #{output_dirname}" |
|
end |
|
|
|
|
|
command = "#{LODA_RUST_EXECUTABLE} arc-size #{TEMP_PATH}" |
|
stdout_and_stderr, status = Open3.capture2e(command) |
|
output = stdout_and_stderr |
|
count_tasks += 1 |
|
|
|
unless status.success? |
|
if output.include?('Cannot predict the output sizes') |
|
output_path2 = output_path.gsub(/[.]json$/, '-cannot-predict.txt') |
|
IO.write(output_path2, stdout_and_stderr) |
|
count_cannot_predict += 1 |
|
next |
|
else |
|
output_path2 = output_path.gsub(/[.]json$/, '-error.txt') |
|
IO.write(output_path2, stdout_and_stderr) |
|
count_other_errors += 1 |
|
next |
|
end |
|
end |
|
json = stdout_and_stderr.strip |
|
predicted_sizes = predicted_sizes(json) |
|
if predicted_sizes != expected_sizes |
|
|
|
output_path2 = output_path.gsub(/[.]json$/, '-bad-prediction.txt') |
|
error_message = stdout_and_stderr + "\n\n--\nThis is a bad prediction!\nPredicted #{predicted_sizes}. But the actual size is #{expected_sizes}" |
|
IO.write(output_path2, error_message) |
|
count_bad_predictions += 1 |
|
next |
|
end |
|
IO.write(output_path, json) |
|
count_ok_predictions += 1 |
|
next |
|
end |
|
end |
|
|
|
File.delete(TEMP_PATH) if File.exist?(TEMP_PATH) |
|
|
|
puts |
|
puts "count_tasks: #{count_tasks} The number of tasks processed." |
|
puts "count_ok_predictions: #{count_ok_predictions} Predictions that matches with the actual data." |
|
puts "count_bad_predictions: #{count_bad_predictions} Predictions that are different than the actual data." |
|
puts "count_cannot_predict: #{count_cannot_predict} Unable to make a prediction. Insufficient data, lack of algorithms for predicting." |
|
puts "count_other_errors: #{count_other_errors} Something went wrong." |
|
puts "count_same_size: #{count_same_size} Number of tasks where input size and output size are the same." |
|
|