Day 2: Red-Nosed Reports
Megathread guidelines
- Keep top level comments as only solutions, if you want to say something other than a solution put it in a new post. (replies to comments can be whatever)
- You can send code in code blocks by using three backticks, the code, and then three backticks or use something such as https://blocks.programming.dev/ if you prefer sending it through a URL
FAQ
- What is this?: Here is a post with a large amount of details: https://programming.dev/post/22323136
- Where do I participate?: https://adventofcode.com/
- Is there a leaderboard for the community?: We have a programming.dev leaderboard with the info on how to join in this post: https://programming.dev/post/6631465
Haskell
This was quite fun! I got a bit distracted trying to rewrite
safe
in point-free style, but I think this version is the most readable. There’s probably a more monadic way of writinglessOne
as well, but I can’t immediately see it.safe xs = any gradual [diffs, negate <$> diffs] where diffs = zipWith (-) (drop 1 xs) xs gradual = all (`elem` [1 .. 3]) lessOne [] = [] lessOne (x : xs) = xs : map (x :) (lessOne xs) main = do input :: [[Int]] <- map (map read . words) . lines <$> readFile "input02" print . length $ filter safe input print . length $ filter (any safe . lessOne) input
Love to see your haskell solutions!
I am so far very amazed with the compactness of your solutions, your
lessOne
is very much mind-Bending. I have never used or seen<$>
before, is it a monadic?
Also I can’t seem to find your logic for this safety condition:
The levels are either all increasing or all decreasing
, did you figure that it wasn’t necessary?
Rust
The function is_sorted_by on Iterators turned out helpful for compactly finding if a report is safe. In part 2 I simply tried the same with each element removed, since all reports are very short.
fn parse(input: String) -> Vec<Vec<i32>> { input.lines() .map(|l| l.split_whitespace().map(|w| w.parse().unwrap()).collect()) .collect() } fn is_safe(report: impl DoubleEndedIterator<Item=i32> + Clone) -> bool { let safety = |a: &i32, b: &i32| (1..=3).contains(&(b - a)); report.clone().is_sorted_by(safety) || report.rev().is_sorted_by(safety) } fn part1(input: String) { let reports = parse(input); let safe = reports.iter().filter(|r| is_safe(r.iter().copied())).count(); println!("{safe}"); } fn is_safe2(report: &[i32]) -> bool { (0..report.len()).any(|i| { // Try with each element removed is_safe(report.iter().enumerate().filter(|(j, _)| *j != i).map(|(_, n)| *n)) }) } fn part2(input: String) { let reports = parse(input); let safe = reports.iter().filter(|r| is_safe2(r)).count(); println!("{safe}"); } util::aoc_main!();
The
is_sorted_by
is a really nice approach. I originally tried using that function thinking that|a, b| a > b
or|a, b| a < b
would cut it but it didn’t end up working. I never thought to handle the check for the step being between 1 and 3 in the callback closure for that though.
Rust
use crate::utils::read_lines; pub fn solution1() { let reports = get_reports(); let safe_reports = reports .filter(|report| report.windows(3).all(window_is_valid)) .count(); println!("Number of safe reports = {safe_reports}"); } pub fn solution2() { let reports = get_reports(); let safe_reports = reports .filter(|report| { (0..report.len()).any(|i| { [&report[0..i], &report[i + 1..]] .concat() .windows(3) .all(window_is_valid) }) }) .count(); println!("Number of safe reports = {safe_reports}"); } fn window_is_valid(window: &[usize]) -> bool { matches!(window[0].abs_diff(window[1]), 1..=3) && matches!(window[1].abs_diff(window[2]), 1..=3) && ((window[0] > window[1] && window[1] > window[2]) || (window[0] < window[1] && window[1] < window[2])) } fn get_reports() -> impl Iterator<Item = Vec<usize>> { read_lines("src/day2/input.txt").map(|line| { line.split_ascii_whitespace() .map(|level| { level .parse() .expect("Reactor level is always valid integer") }) .collect() }) }
Definitely trickier than yesterday’s. I feel like the
windows
solution isn’t the best, but it was what came to mind and ended up working for me.