* add GQA for llama2 * fix model conversion * fix lint & remove dev log * update news * minor * fix allocation size * fix split_dim for w_qkv.bias
* change target tritonfastertransformerbackend to tritonturbomindbackend * install targets to backends/turbomind * changge model_dir
* build turbomind * change namespace fastertransformer to turbomind * change logger name