Researchers from the University of Maryland, Lawrence Livermore, Columbia and TogetherAI have developed a training technique that triples LLM inference speed without auxiliary models or infrastructure ...
Abstract: Transformer models suffer from unaffordable high memory consumption when the sequence is long and standard self-attention is utilized. We developed a sequence parallelism scheme called ...
A Sequence Fine-Tuning Strategy Based Evolutionary Algorithm for Solving Project Scheduling Problems
Abstract: As artificial intelligence technology advances, the industrial landscape has been gradually transitioning from man-ual labor to automated manufacturing processes. This shift has highlighted ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results