java流stream中的Collectors中groupingBy源码笔记

    /**
     * Returns a {@code Collector} implementing a cascaded "group by" operation
     * on input elements of type {@code T}, grouping elements according to a
     * classification function, and then performing a reduction operation on
     * the values associated with a given key using the specified downstream
     * {@code Collector}.  The {@code Map} produced by the Collector is created
     * with the supplied factory function.
     返回一个可级联的实现了分组功能的收集器,这个收集器根据T类型的输入参数,通过分类函数进行分类,然后使用指定的下游收集器执行一个汇聚操作。这个收集器的结果容器Map,由给定的工厂方法提供。
     *
     * <p>The classification function maps elements to some key type {@code K}.
     * The downstream collector operates on elements of type {@code T} and
     * produces a result of type {@code D}. The resulting collector produces a
     * {@code Map<K, D>}.
     分类方法将T类型的输入元素映射为K类型的key,作为结果map的key,下游收集器将T类型的输入元素转化为D类型的结果元素,最终结果收集器生产出Map<K,D>的结果。
     *
     * <p>For example, to compute the set of last names of people in each city,
     * where the city names are sorted:
     例子,收集人民群众的姓氏,结果根据城市分组。
     * <pre>{@code
     * Map<City, Set<String>> namesByCity
     *   = people.stream().collect(
     *     groupingBy(Person::getCity,对应分类函数classifier
     *                TreeMap::new,对应结果容器工厂mapFactory
     *                mapping(Person::getLastName, 对应下游收集器downstream
     *                        toSet())));
     * }</pre>
     *groupingBy对应结果收集器,是最终的收集器。
     * @implNote
     * The returned {@code Collector} is not concurrent.  For parallel stream
     * pipelines, the {@code combiner} function operates by merging the keys
     * from one map into another, which can be an expensive operation.  If
     * preservation of the order in which elements are presented to the downstream
     * collector is not required, using {@link #groupingByConcurrent(Function, Supplier, Collector)}
     * may offer better parallel performance.
     返回的收集器不是并发的,对于并发流来说,组合器合并map的操作可能会很耗性能。
     如果不需要保持元素在流中的顺序,推荐使用groupingByConcurrent,这可能要比使用parallel stream的性能更好。
     
     * @param <T> the type of the input elements T:输入元素的类型
     * @param <K> the type of the keys K:结果map中的key类型。
     * @param <A> the intermediate accumulation type of the downstream collector
     * @param <D> the result type of the downstream reduction
     * @param <M> the type of the resulting {@code Map}
     * @param classifier a classifier function mapping input elements to keys
     * @param downstream a {@code Collector} implementing the downstream reduction
     * @param mapFactory a supplier providing a new empty {@code Map}
     *                   into which the results will be inserted
     * @return a {@code Collector} implementing the cascaded group-by operation
     *
     * @see #groupingBy(Function, Collector)
     * @see #groupingBy(Function)
     * @see #groupingByConcurrent(Function, Supplier, Collector)
     T:输入元素的类型。
     K:结果map中的key类型。
     A: 下游收集器的累加器的容器类型(累加器的第一个参数)。
     D: 下游收集器的结果类型。当下游收集器没有finisher的时候,A和D是直接相等的。A强转为D。
     M: 最终结果类型,即Map<K,D>
     最后返回一个实现了可级联分组的收集器。
     
     这个方法总体来讲,就是给一个分组器,一个最终类型的生产者,一个收集器,根据这三个参数,来改造出一个能分组的收集器。
     */
    public static <T, K, D, A, M extends Map<K, D>> //注意这里有5个参数类型
    Collector<T, ?, M> groupingBy(Function<? super T, ? extends K> classifier,
                                  Supplier<M> mapFactory,
                                  Collector<? super T, A, D> downstream) {
        Supplier<A> downstreamSupplier = downstream.supplier();
        BiConsumer<A, ? super T> downstreamAccumulator = downstream.accumulator();
        BiConsumer<Map<K, A>, T> accumulator = (m, t) -> {
            // 根据分类器得到的值,最为最终map中的键
            K key = Objects.requireNonNull(classifier.apply(t), "element cannot be mapped to a null key");
            // 得到一下游收集器的生产者生产的容器,最为最终map中的值。
            A container = m.computeIfAbsent(key, k -> downstreamSupplier.get());
            // 消费这两个参数,进行累加操作,相当于修改了下游收集器的收集过程,让其成为最终收集器的累加器,累积出最终收集器需要的中间结果。
            downstreamAccumulator.accept(container, t);
        };
        // 传入下游收集器的合并器,得到一个新的合并器,合并器合出来的值是经过改造的累加器的结果,所以是合出的最终类型Map<K,A>
        BinaryOperator<Map<K, A>> merger = Collectors.<K, A, Map<K, A>>mapMerger(downstream.combiner());
        // 将Map<K, D>类型的mapFactory强转为Map<K, A>类型,这其中包含了A到D的强转。
        @SuppressWarnings("unchecked")
        Supplier<Map<K, A>> mangledFactory = (Supplier<Map<K, A>>) mapFactory;
        // 如果集合特性包含IDENTITY_FINISH,说明下游收集器的中间结果就是最终结果,不用再处理finisher
        if (downstream.characteristics().contains(Collector.Characteristics.IDENTITY_FINISH)) {
            return new CollectorImpl<>(mangledFactory, accumulator, merger, CH_ID);
        } else {
            //将下游收集器的finisher强转为输入A,输出A的finisher(限定类型)
            @SuppressWarnings("unchecked")
            Function<A, A> downstreamFinisher = (Function<A, A>) downstream.finisher();
            // 用强转好的finisher,处理所有元素,这时的元素是一个一个的map(前面合并的)
            // intermediate代表一个map,将每个map都用改过的finisher处理一下,得到Map<K,A>类型,再强转一下,将Map<K,A>强转为Map<K,D>
            Function<Map<K, A>, M> finisher = intermediate -> {
                // 这里replace的只是value,将value处理成A类型
                intermediate.replaceAll((k, v) -> downstreamFinisher.apply(v));
                @SuppressWarnings("unchecked")
                M castResult = (M) intermediate;
                return castResult;
            };
            return new CollectorImpl<>(mangledFactory, accumulator, merger, finisher, CH_NOID);
        }
    }


    /**
     * {@code BinaryOperator<Map>} that merges the contents of its right
     * argument into its left argument, using the provided merge function to
     * handle duplicate keys.
     将右边的参数合并到左边
     *
     * @param <K> type of the map keys
     * @param <V> type of the map values
     * @param <M> type of the map
     * @param mergeFunction A merge function suitable for
     * {@link Map#merge(Object, Object, BiFunction) Map.merge()}
     * @return a merge function for two maps
     */
    private static <K, V, M extends Map<K,V>>
    BinaryOperator<M> mapMerger(BinaryOperator<V> mergeFunction) {
        return (m1, m2) -> {
            for (Map.Entry<K,V> e : m2.entrySet())
                // 如果左边的map里面,左边map中没有右边合过来的key对应的值,就用右边合过来的值,
                // ,如果有值,就使用合并器算出来的值,确保不冲突。
                m1.merge(e.getKey(), e.getValue(), mergeFunction);
            return m1;
        };
    }
原文地址:https://www.cnblogs.com/Lothlorien/p/12142461.html