JGibbLDA的输出文件

五个输出文件:


  • model-final.towords
    每个topic下面的words和words分布,并按分布排序:
1Topic 0th:
	bill 0.005843543826578699
	lai 0.003958529688972668
	seventh 0.0020735155513666352
	immedi 0.0020735155513666352
	anaheim 0.0020735155513666352
	concern 0.0020735155513666352
	month 0.0020735155513666352
	american 0.0020735155513666352
	decision 0.0020735155513666352
	risk 0.0020735155513666352
	demis 0.0020735155513666352
	maintain 0.0020735155513666352
	rose 0.0020735155513666352
Topic 1th:
	photo 0.0040191387559808615
	seri 0.0040191387559808615
	fashion 0.0040191387559808615
	subject 0.0021052631578947372
	left 0.0021052631578947372
	briefli 0.0021052631578947372
	discov 0.0021052631578947372
	hick 0.0021052631578947372
	line 0.0021052631578947372
	lip 0.0021052631578947372
	whom 0.0021052631578947372
	beij 0.0021052631578947372
	man 0.0021052631578947372
	xiaop 0.0021052631578947372
	highli 0.0021052631578947372
	california 0.0021052631578947372
Topic 2th:
	enjoi 0.0039069767441860465
	fan 0.0039069767441860465
	shuttle 0.0039069767441860465
	return 0.002046511627906977
	little 0.002046511627906977
	plai 0.002046511627906977
	songwrit 0.002046511627906977
	sai 0.002046511627906977
	debat 0.002046511627906977
	depart 0.002046511627906977
	disagre 0.002046511627906977
	iii 0.002046511627906977
	jan 0.002046511627906977
	mere 0.002046511627906977
	relat 0.002046511627906977
	stress 0.002046511627906977
  • model-final.phi
    topic -words distributions
    是上面towords文件的原始数据文件,未排序,是数组里面直接保存的数据.

  • model-final.theta
    每个doc下的topic分布,无排序

  • model-final.assign
    每个word所属的topic代号:
    0:35
    0代表word的id,35代表第36个topic

0:35 1:30 1:52 2:11 2:11 2:11 2:11 3:52 4:11 5:26 6:52 7:38 8:98 9:93 10:11 11:11 12:33 12:52 12:11 12:11 13:11 14:52 14:52 15:12 16:11 16:11 
7:11 18:11 18:11 19:13 20:52 20:52 21:66 22:93 23:21 24:98 25:81 26:11 27:11 28:7 29:52 30:11 31:11 31:11 31:11 32:11 32:11 33:11 34:57 34:52 34:52 35:11 36:25 37:11 37:11 37:11 38:32 39:11 40:1 40:52 41:52 42:52 42:52 43:52 43:52 44:11 45:14 46:11 47:11 48:38 49:11 50:36 51:52 52:12 
53:65 54:83 55:10 56:27 57:4 58:52 59:52 60:24 60:35 61:63 62:83 14:52 14:52 63:89 63:83 64:5 65:43 66:52 20:52 20:52 67:81 68:41 69:65 70:41 71:41 72:83 
73:88 74:83 75:24 76:83 77:73 78:10 79:24 80:5 81:52 40:52 40:52 82:52 42:52 42:52 43:52 43:52 51:52 
  • model-final.others
    一些参数
alpha=0.5
beta=0.1
ntopics=100
ndocs=100
nwords=4815
liters=999

根据结果文件,发现训练集和测试集的结果文件类型数量相同,并且类型也相同.所以没必要把estimate和inference拆开写.
另外inference的测试过程也是一种训练,只不过训练时要参考原有训练集训练的结果,这样会使训练更准确.
测试集new的model叫newModel,如果测试集的数据量比原有测试集还要大,这newModel可以代替原有的训练集.


问题:

测试集的训练过程是以自身为主,还是以原有训练集为主.要弄清楚这一点.


下一步计划:

1.下面的工作是继续分析测试集的内容,并且重新分析这个代码如何完成.
2.给源代码添加文件,让各种结果可视化.
3.中文lda.

原文地址:https://www.cnblogs.com/cyno/p/4451809.html