X-Git-Url: https://git.xandkar.net/?p=dups.git;a=blobdiff_plain;f=dups.ml;h=2521b97261fbc2fa441283aeefc1fb5a78e27c54;hp=6afba6a893f8d90e5ca9362d5d924ff3078d1ac9;hb=b0c3b3f085147e4228cf17a965723feb041b275f;hpb=217f89128605668889d84e576e595248e91b11c2 diff --git a/dups.ml b/dups.ml index 6afba6a..2521b97 100644 --- a/dups.ml +++ b/dups.ml @@ -283,51 +283,48 @@ end = struct end let lord t ~njobs ~vassals ~ic ~ocs = - eprintf "[debug] [lord] started\n%!"; let active_vassals = ref njobs in let results = Queue.create () in - let rec dispatch () = - match Ipc.recv ic with - | ((Exiting i) : ('input, 'output) msg_from_vassal) -> + let rec loop () = + match ((Ipc.recv ic) : ('input, 'output) msg_from_vassal) with + | Exiting i -> close_out ocs.(i); decr active_vassals; - if !active_vassals = 0 then - () - else - dispatch () - | ((Ready i) : ('input, 'output) msg_from_vassal) -> + if !active_vassals = 0 then () else loop () + | Ready i -> Ipc.send ocs.(i) (Job (next t)); - dispatch () - | ((Result (i, result)) : ('input, 'output) msg_from_vassal) -> + loop () + | Result (i, result) -> Queue.add result results; Ipc.send ocs.(i) (Job (next t)); - dispatch () + loop () in let rec wait = function - | [] -> () + | [] -> + () | vassals -> let pid, _process_status = Unix.wait () in (* TODO: handle process_status *) wait (List.filter vassals ~f:(fun p -> p <> pid)) in - dispatch (); + loop (); close_in ic; wait vassals; of_queue results let vassal i ~f ~vassal_pipe_r ~lord_pipe_w = - eprintf "[debug] [vassal %d] started\n%!" i; let ic = Unix.in_channel_of_descr vassal_pipe_r in let oc = Unix.out_channel_of_descr lord_pipe_w in - let rec work msg = - Ipc.send oc msg; - match Ipc.recv ic with - | (Job (Some x) : 'input msg_from_lord) -> - work (Result (i, (x, f x))) - | (Job None : 'input msg_from_lord) -> + let rec loop () = + match (Ipc.recv ic : 'input msg_from_lord) with + | Job (Some x) -> + Ipc.send oc (Result (i, (x, f x))); + loop () + | Job None -> Ipc.send oc (Exiting i) in - work (Ready i); + Ipc.send oc (Ready i); + loop (); close_in ic; close_out oc; exit 0 @@ -357,16 +354,27 @@ end = struct end module In_channel : sig - val lines : in_channel -> string Stream.t + val lines : ?delim_null:bool -> in_channel -> string Stream.t end = struct - let lines ic = - Stream.create (fun () -> - match input_line ic with - | exception End_of_file -> - None - | line -> - Some line - ) + let read_until_newline ic () = + match input_line ic with + | exception End_of_file -> + None + | line -> + Some line + + let read_until_null ic = + let lexbuf = Lexing.from_channel ic in + fun () -> Input_delim.by_null lexbuf + + let lines ?(delim_null=false) ic = + let reader = + if delim_null then + read_until_null ic + else + read_until_newline ic + in + Stream.create reader end module File : sig @@ -381,6 +389,8 @@ module File : sig val lookup : string Stream.t -> t Stream.t (** Lookup file info for given paths *) + val head : t -> len:int -> metrics:M.t -> string + val filter_out_unique_sizes : t Stream.t -> metrics:M.t -> t Stream.t val filter_out_unique_heads : t Stream.t -> len:int -> metrics:M.t -> t Stream.t end = struct @@ -444,7 +454,8 @@ end = struct ~group:(fun {size; _} -> size) ~handle_singleton:(fun (size, _, _) -> M.file_unique_size metrics ~size) - let head path ~len ~metrics = + let head {path; _} ~len ~metrics = + M.file_sampled metrics; let buf = Bytes.make len ' ' in let ic = open_in_bin path in let rec read pos len = @@ -467,10 +478,7 @@ end = struct let filter_out_unique_heads files ~len ~metrics = filter_out_singletons files - ~group:(fun {path; _} -> - M.file_sampled metrics; - head path ~len ~metrics - ) + ~group:(head ~len ~metrics) ~handle_singleton:(fun (_, _, files) -> let {size; _} = List.hd files in (* Guaranteed non-empty *) M.file_unique_sample metrics ~size @@ -491,13 +499,14 @@ type opt = ; ignore : string -> bool ; sample : int ; njobs : int + ; delim_null : bool } -let make_input_stream input ignore ~metrics = +let make_input_stream input ignore ~metrics ~delim_null = let input = match input with | Stdin -> - File.lookup (In_channel.lines stdin) + File.lookup (In_channel.lines stdin ~delim_null) | Directories paths -> let paths = StrSet.elements (StrSet.of_list paths) in Stream.concat (List.map paths ~f:File.find) @@ -535,12 +544,12 @@ let time_wall () = let time_proc () = Sys.time () -let main {input; output; ignore; sample = sample_len; njobs} = +let main {input; output; ignore; sample = sample_len; njobs; delim_null} = let wt0_all = time_wall () in let pt0_all = time_proc () in let metrics = M.init () in let output = make_output_fun output in - let input = make_input_stream input ignore ~metrics in + let input = make_input_stream input ignore ~metrics ~delim_null in (* TODO: Make a nice(r) abstraction to re-assemble pieces in the pipeline: * * from input to files_by_size @@ -563,7 +572,20 @@ let main {input; output; ignore; sample = sample_len; njobs} = let wt0_group_by_sample = wt1_group_by_size in let pt0_group_by_sample = pt1_group_by_size in eprintf "[debug] filtering-out files with unique heads\n%!"; - let files = File.filter_out_unique_heads files ~len:sample_len ~metrics in + let files = + if njobs > 1 then begin + let q = Queue.create () in + files + |> Stream.bag_map ~njobs ~f:(File.head ~len:sample_len ~metrics) + |> Stream.group_by ~f:snd + |> Stream.map ~f:(fun (d, n, pairs) -> (d, n, List.map pairs ~f:fst)) + |> Stream.filter ~f:(fun (_, n, _) -> n > 1) + |> Stream.iter ~f:(fun (_, _, fs) -> List.iter fs ~f:(fun f -> Queue.add f q)) + ; + Stream.of_queue q + end else + File.filter_out_unique_heads files ~len:sample_len ~metrics + in let pt1_group_by_sample = time_proc () in let wt1_group_by_sample = time_wall () in @@ -572,10 +594,10 @@ let main {input; output; ignore; sample = sample_len; njobs} = eprintf "[debug] hashing\n%!"; let groups = if njobs > 1 then - let digests = + let with_digests = Stream.bag_map files ~njobs ~f:(fun {File.path; _} -> Digest.file path) in - Stream.map (Stream.group_by digests ~f:(fun (_, d) -> d)) ~f:( + Stream.map (Stream.group_by with_digests ~f:snd) ~f:( fun (digest, n, file_digest_pairs) -> let files = List.map file_digest_pairs ~f:(fun (file, _) -> @@ -597,9 +619,10 @@ let main {input; output; ignore; sample = sample_len; njobs} = eprintf "[debug] reporting\n%!"; Stream.iter groups ~f:(fun (d, n, files) -> M.digest metrics; - if n > 1 then + if n > 1 then begin M.redundant_data metrics ~size:(n * (List.hd files).File.size); output d n files + end ); let pt1_all = time_proc () in @@ -632,7 +655,8 @@ let get_opt () : opt = let output = ref Stdout in let ignore = ref (fun _ -> false) in let sample = ref 512 in - let njobs = ref 8 in + let njobs = ref 6 in + let input_delim_null = ref false in let spec = [ ( "-out" , Arg.String (fun path -> @@ -656,6 +680,16 @@ let get_opt () : opt = , Arg.Set_int njobs , (sprintf " Number of parallel jobs. Default: %d" !njobs) ) + ; ( "-0" + , Arg.Set input_delim_null + , ( sprintf + ( " Delimit input paths by null character instead of a newline." + ^^" Meaningful only when reading candidate paths from stdin." + ^^" Default: %B" + ) + !input_delim_null + ) + ) ] in Arg.parse @@ -679,6 +713,7 @@ let get_opt () : opt = ; ignore = !ignore ; sample = !sample ; njobs = !njobs + ; delim_null = !input_delim_null } let () =